error577 commited on
Commit
0d18af4
·
verified ·
1 Parent(s): b9aaa8b

End of training

Browse files
README.md CHANGED
@@ -65,7 +65,7 @@ lora_model_dir: null
65
  lora_r: 32
66
  lora_target_linear: true
67
  lr_scheduler: cosine
68
- max_steps: 100
69
  micro_batch_size: 1
70
  mlflow_experiment_name: /tmp/0c389fe694d49e2c_train_data.json
71
  model_type: AutoModelForCausalLM
@@ -83,7 +83,7 @@ tf32: false
83
  tokenizer_type: AutoTokenizer
84
  train_on_inputs: false
85
  trust_remote_code: true
86
- val_set_size: 0.02
87
  wandb_entity: null
88
  wandb_mode: online
89
  wandb_name: 8258b119-6362-4754-9e34-cdd215dba50b
@@ -130,17 +130,17 @@ The following hyperparameters were used during training:
130
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
131
  - lr_scheduler_type: cosine
132
  - lr_scheduler_warmup_steps: 10
133
- - training_steps: 100
134
 
135
  ### Training results
136
 
137
  | Training Loss | Epoch | Step | Validation Loss |
138
  |:-------------:|:------:|:----:|:---------------:|
139
- | 4.5019 | 0.0004 | 1 | nan |
140
- | 2.3903 | 0.0100 | 25 | nan |
141
- | 2.9039 | 0.0199 | 50 | nan |
142
- | 2.4618 | 0.0299 | 75 | nan |
143
- | 1.779 | 0.0398 | 100 | nan |
144
 
145
 
146
  ### Framework versions
 
65
  lora_r: 32
66
  lora_target_linear: true
67
  lr_scheduler: cosine
68
+ max_steps: 200
69
  micro_batch_size: 1
70
  mlflow_experiment_name: /tmp/0c389fe694d49e2c_train_data.json
71
  model_type: AutoModelForCausalLM
 
83
  tokenizer_type: AutoTokenizer
84
  train_on_inputs: false
85
  trust_remote_code: true
86
+ val_set_size: 0.1
87
  wandb_entity: null
88
  wandb_mode: online
89
  wandb_name: 8258b119-6362-4754-9e34-cdd215dba50b
 
130
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
131
  - lr_scheduler_type: cosine
132
  - lr_scheduler_warmup_steps: 10
133
+ - training_steps: 200
134
 
135
  ### Training results
136
 
137
  | Training Loss | Epoch | Step | Validation Loss |
138
  |:-------------:|:------:|:----:|:---------------:|
139
+ | 4.2027 | 0.0004 | 1 | nan |
140
+ | 2.7534 | 0.0217 | 50 | nan |
141
+ | 1.7026 | 0.0434 | 100 | nan |
142
+ | 2.3272 | 0.0650 | 150 | nan |
143
+ | 2.5038 | 0.0867 | 200 | nan |
144
 
145
 
146
  ### Framework versions
adapter_config.json CHANGED
@@ -20,11 +20,11 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "up_proj",
24
- "down_proj",
25
  "gate_proj",
26
- "o_proj",
27
  "v_proj",
 
 
 
28
  "q_proj",
29
  "k_proj"
30
  ],
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "gate_proj",
 
24
  "v_proj",
25
+ "down_proj",
26
+ "up_proj",
27
+ "o_proj",
28
  "q_proj",
29
  "k_proj"
30
  ],
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:479d3eef82ce226f70d01e323bbec796aa6fa19cd7fd64dd349d734b9f409782
3
  size 335706186
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:316e70f9b3f4f0a73b768477122e631b34b1aab87bef3c37e176ad310169a558
3
  size 335706186
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4bae772233d95878af6c79ca0786e029c5b928ca1ddac881139ef2dd842e35f
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c7999ded8de52ab72cb7097e1002a337b6415095697681252714d881491b44b
3
  size 335604696
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e06327c333702c90c3c2773755c7143c276c333539c2f472aeb955071203f580
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a0f802361e7d4ff965c2243609680b1fcf327f5a2f5e825bea35cfb3727ce91
3
  size 6776