enumerator commited on
Commit
4697660
·
verified ·
1 Parent(s): ec98203

Training in progress, step 50

Browse files
README.md CHANGED
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/erdenkalilaeyev-bah-e-ehir-cyprus-university/huggingface/runs/pddfjaby)
31
 
32
 
33
  This model was trained with SFT.
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/erdenkalilaeyev-bah-e-ehir-cyprus-university/huggingface/runs/bsu7eh33)
31
 
32
 
33
  This model was trained with SFT.
adapter_config.json CHANGED
@@ -13,7 +13,7 @@
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
- "lora_alpha": 16,
17
  "lora_bias": false,
18
  "lora_dropout": 0.05,
19
  "megatron_config": null,
@@ -23,20 +23,12 @@
23
  "embed_tokens"
24
  ],
25
  "peft_type": "LORA",
26
- "r": 32,
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
30
- "o_proj",
31
- "k_proj",
32
  "v_proj",
33
- "fc2",
34
- "q_proj",
35
- "gate_proj",
36
- "fc1",
37
- "up_proj",
38
- "down_proj",
39
- "out_proj"
40
  ],
41
  "task_type": "CAUSAL_LM",
42
  "trainable_token_indices": null,
 
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
+ "lora_alpha": 8,
17
  "lora_bias": false,
18
  "lora_dropout": 0.05,
19
  "megatron_config": null,
 
23
  "embed_tokens"
24
  ],
25
  "peft_type": "LORA",
26
+ "r": 8,
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
 
 
30
  "v_proj",
31
+ "q_proj"
 
 
 
 
 
 
32
  ],
33
  "task_type": "CAUSAL_LM",
34
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2dcd26aadf24f808813c023ca32bea2b31a427e55ce3bd78d4b712085b07376b
3
- size 2993118096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e198a48c4a129ecd196a3dfa192389e899b093f1f6ccb804438dbb16a01c2622
3
+ size 2697942384
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a305404e6f9dc8297963c448c1d49430687151ccbc8db599f96037eb3943bafb
3
  size 33384822
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da0b388d2d5c9938e523ade0f9a37b4b8fb1b621c8bfd42edbb506309e93e711
3
  size 33384822
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2302e944cf78fcb479b1349e190717a509833854ebcf407a67a33cdd9e6f4bd6
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1790ce2ff92522986951769c96444b9a6b6a20279d5251b0f21d5949f55416b0
3
  size 5752