nanaseven commited on
Commit
758ac3a
·
verified ·
1 Parent(s): 3eed61a

End of training

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: Gensyn/Qwen2.5-0.5B-Instruct
3
  library_name: transformers
4
  model_name: gensyn-checkpoints-alert_pawing_anaconda
5
  tags:
@@ -8,13 +8,14 @@ tags:
8
  - grpo
9
  - gensyn
10
  - I am alert pawing anaconda
 
11
  - trl
12
  licence: license
13
  ---
14
 
15
  # Model Card for gensyn-checkpoints-alert_pawing_anaconda
16
 
17
- This model is a fine-tuned version of [Gensyn/Qwen2.5-0.5B-Instruct](https://huggingface.co/Gensyn/Qwen2.5-0.5B-Instruct).
18
  It has been trained using [TRL](https://github.com/huggingface/trl).
19
 
20
  ## Quick start
@@ -39,8 +40,8 @@ This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing
39
 
40
  - TRL: 0.15.2
41
  - Transformers: 4.51.3
42
- - Pytorch: 2.5.1
43
- - Datasets: 3.5.0
44
  - Tokenizers: 0.21.1
45
 
46
  ## Citations
 
1
  ---
2
+ base_model: Gensyn/Qwen2.5-1.5B-Instruct
3
  library_name: transformers
4
  model_name: gensyn-checkpoints-alert_pawing_anaconda
5
  tags:
 
8
  - grpo
9
  - gensyn
10
  - I am alert pawing anaconda
11
+ - unsloth
12
  - trl
13
  licence: license
14
  ---
15
 
16
  # Model Card for gensyn-checkpoints-alert_pawing_anaconda
17
 
18
+ This model is a fine-tuned version of [Gensyn/Qwen2.5-1.5B-Instruct](https://huggingface.co/Gensyn/Qwen2.5-1.5B-Instruct).
19
  It has been trained using [TRL](https://github.com/huggingface/trl).
20
 
21
  ## Quick start
 
40
 
41
  - TRL: 0.15.2
42
  - Transformers: 4.51.3
43
+ - Pytorch: 2.6.0
44
+ - Datasets: 3.6.0
45
  - Tokenizers: 0.21.1
46
 
47
  ## Citations
adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Gensyn/Qwen2.5-1.5B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "v_proj",
28
+ "k_proj",
29
+ "down_proj",
30
+ "gate_proj",
31
+ "q_proj",
32
+ "up_proj",
33
+ "o_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a493fcd4f8490e6da17c81266ac2321fb8c0b900cea4e713346926af3f7089b
3
+ size 73911112
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 2.26292759180069e-06,
4
- "train_runtime": 300.3933,
5
- "train_samples": 28,
6
- "train_samples_per_second": 1.065,
7
- "train_steps_per_second": 0.067
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 7.334376395995435e-07,
4
+ "train_runtime": 1213.9576,
5
+ "train_samples": 15,
6
+ "train_samples_per_second": 0.264,
7
+ "train_steps_per_second": 0.016
8
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 2.26292759180069e-06,
4
- "train_runtime": 300.3933,
5
- "train_samples": 28,
6
- "train_samples_per_second": 1.065,
7
- "train_steps_per_second": 0.067
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 7.334376395995435e-07,
4
+ "train_runtime": 1213.9576,
5
+ "train_samples": 15,
6
+ "train_samples_per_second": 0.264,
7
+ "train_steps_per_second": 0.016
8
  }
trainer_state.json CHANGED
@@ -10,203 +10,203 @@
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "completion_length": 409.5625,
14
- "epoch": 0.5714285714285714,
15
- "grad_norm": 7.3488874435424805,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
  "loss": -0.0,
19
- "reward": 1.911701524630189,
20
- "reward_std": 1.9608921660110354,
21
- "rewards/concensus_correctness_reward_func": 0.6924999989569187,
22
- "rewards/consensus_reward_func": 0.1875,
23
  "rewards/cumulative_reward_2": 0.0,
24
- "rewards/final_correctness_reward_func": 0.125,
25
- "rewards/question_recreation_reward_func": 0.502670262940228,
26
  "rewards/soft_format_reward_func": 0.0,
27
- "rewards/strict_format_reward_func": 0.0,
28
- "rewards/xmlcount_reward_func": 0.40403125062584877,
29
  "step": 2
30
  },
31
  {
32
- "completion_length": 328.0833333333333,
33
  "epoch": 1.0,
34
- "grad_norm": 2.5143661499023438,
35
- "kl": 0.000962535836151801,
36
  "learning_rate": 4.864543104251586e-07,
37
  "loss": 0.0,
38
- "reward": 1.6085879653692245,
39
- "reward_std": 0.4928823659817378,
40
- "rewards/concensus_correctness_reward_func": 0.07999999821186066,
41
- "rewards/consensus_reward_func": 0.0,
42
  "rewards/cumulative_reward_2": 0.0,
43
- "rewards/final_correctness_reward_func": 0.16666666666666666,
44
- "rewards/question_recreation_reward_func": 0.5609629477063814,
45
- "rewards/soft_format_reward_func": 0.0,
46
  "rewards/strict_format_reward_func": 0.0,
47
- "rewards/xmlcount_reward_func": 0.8009583279490471,
48
  "step": 4
49
  },
50
  {
51
- "completion_length": 399.75,
52
- "epoch": 1.5714285714285714,
53
- "grad_norm": 5.2374091148376465,
54
- "kl": 0.0016545295457035536,
55
  "learning_rate": 4.472851273490984e-07,
56
  "loss": 0.0,
57
- "reward": 1.8188971318304539,
58
- "reward_std": 1.099338564556092,
59
- "rewards/concensus_correctness_reward_func": 0.6851250007748604,
60
- "rewards/consensus_reward_func": 0.0,
61
  "rewards/cumulative_reward_2": 0.0,
62
- "rewards/final_correctness_reward_func": 0.0625,
63
- "rewards/question_recreation_reward_func": 0.5636470918543637,
64
  "rewards/soft_format_reward_func": 0.0,
65
- "rewards/strict_format_reward_func": 0.0,
66
- "rewards/xmlcount_reward_func": 0.5076250061392784,
67
  "step": 6
68
  },
69
  {
70
- "completion_length": 369.0833333333333,
71
  "epoch": 2.0,
72
- "grad_norm": 2.376291275024414,
73
- "kl": 0.0015570806475201,
74
  "learning_rate": 3.867370395306068e-07,
75
  "loss": 0.0,
76
- "reward": 2.6855219999949136,
77
- "reward_std": 2.0642324524621167,
78
- "rewards/concensus_correctness_reward_func": 0.9224166671435038,
79
- "rewards/consensus_reward_func": 0.25,
80
  "rewards/cumulative_reward_2": 0.0,
81
- "rewards/final_correctness_reward_func": 0.25,
82
- "rewards/question_recreation_reward_func": 0.6277303335567316,
83
  "rewards/soft_format_reward_func": 0.0,
84
- "rewards/strict_format_reward_func": 0.0,
85
- "rewards/xmlcount_reward_func": 0.6353750017782053,
86
  "step": 8
87
  },
88
  {
89
- "completion_length": 424.96875,
90
- "epoch": 2.571428571428571,
91
- "grad_norm": 3.9620633125305176,
92
- "kl": 0.002315260087925708,
93
  "learning_rate": 3.1137137178519977e-07,
94
  "loss": 0.0,
95
- "reward": 1.3290682332590222,
96
- "reward_std": 0.747250135987997,
97
- "rewards/concensus_correctness_reward_func": 0.07956250011920929,
98
- "rewards/consensus_reward_func": 0.125,
99
  "rewards/cumulative_reward_2": 0.0,
100
- "rewards/final_correctness_reward_func": 0.1875,
101
- "rewards/question_recreation_reward_func": 0.41284948121756315,
102
  "rewards/soft_format_reward_func": 0.0,
103
- "rewards/strict_format_reward_func": 0.0,
104
- "rewards/xmlcount_reward_func": 0.524156260304153,
105
  "step": 10
106
  },
107
  {
108
- "completion_length": 423.0833333333333,
109
  "epoch": 3.0,
110
- "grad_norm": 2.795499563217163,
111
- "kl": 0.003039477189304307,
112
  "learning_rate": 2.2935516363191693e-07,
113
  "loss": 0.0,
114
- "reward": 1.7231739660104115,
115
- "reward_std": 1.0104351732879877,
116
- "rewards/concensus_correctness_reward_func": 0.16308333228031793,
117
- "rewards/consensus_reward_func": 0.4166666666666667,
118
  "rewards/cumulative_reward_2": 0.0,
119
- "rewards/final_correctness_reward_func": 0.25,
120
- "rewards/question_recreation_reward_func": 0.5202989131212234,
121
  "rewards/soft_format_reward_func": 0.0,
122
- "rewards/strict_format_reward_func": 0.0,
123
- "rewards/xmlcount_reward_func": 0.37312500551342964,
124
  "step": 12
125
  },
126
  {
127
- "completion_length": 377.09375,
128
- "epoch": 3.571428571428571,
129
- "grad_norm": 13.31064510345459,
130
- "kl": 0.0036816853971686214,
131
  "learning_rate": 1.4957614383675767e-07,
132
  "loss": 0.0,
133
- "reward": 1.623508368153125,
134
- "reward_std": 0.8322483664378524,
135
- "rewards/concensus_correctness_reward_func": 0.10193750099278986,
136
- "rewards/consensus_reward_func": 0.125,
137
  "rewards/cumulative_reward_2": 0.0,
138
- "rewards/final_correctness_reward_func": 0.1875,
139
- "rewards/question_recreation_reward_func": 0.5396646414883435,
140
  "rewards/soft_format_reward_func": 0.0,
141
- "rewards/strict_format_reward_func": 0.0,
142
- "rewards/xmlcount_reward_func": 0.669406246393919,
143
  "step": 14
144
  },
145
  {
146
- "completion_length": 344.9583333333333,
147
  "epoch": 4.0,
148
- "grad_norm": 4.11659574508667,
149
- "kl": 0.005116298232072343,
150
  "learning_rate": 8.067960709356478e-08,
151
  "loss": 0.0,
152
- "reward": 1.2915391772985458,
153
- "reward_std": 0.7148025852317611,
154
- "rewards/concensus_correctness_reward_func": 0.03758333375056585,
155
- "rewards/consensus_reward_func": 0.08333333333333333,
156
  "rewards/cumulative_reward_2": 0.0,
157
- "rewards/final_correctness_reward_func": 0.16666666666666666,
158
- "rewards/question_recreation_reward_func": 0.4621641853203376,
159
  "rewards/soft_format_reward_func": 0.0,
160
  "rewards/strict_format_reward_func": 0.0,
161
- "rewards/xmlcount_reward_func": 0.5417916551232338,
162
  "step": 16
163
  },
164
  {
165
- "completion_length": 501.90625,
166
- "epoch": 4.571428571428571,
167
- "grad_norm": 4.820127487182617,
168
- "kl": 0.0034084110739058815,
169
  "learning_rate": 3.013156219837776e-08,
170
  "loss": 0.0,
171
- "reward": 1.9342150054872036,
172
- "reward_std": 2.3277871979225893,
173
- "rewards/concensus_correctness_reward_func": 0.6885625012218952,
174
- "rewards/consensus_reward_func": 0.25,
175
  "rewards/cumulative_reward_2": 0.0,
176
- "rewards/final_correctness_reward_func": 0.25,
177
- "rewards/question_recreation_reward_func": 0.5643087650532834,
178
  "rewards/soft_format_reward_func": 0.0,
179
  "rewards/strict_format_reward_func": 0.0,
180
- "rewards/xmlcount_reward_func": 0.18134376220405102,
181
  "step": 18
182
  },
183
  {
184
- "completion_length": 415.9166666666667,
185
  "epoch": 5.0,
186
- "grad_norm": 3.886725902557373,
187
- "kl": 0.003820983639646632,
188
  "learning_rate": 3.4096741493194193e-09,
189
  "loss": 0.0,
190
- "reward": 1.0419001132249832,
191
- "reward_std": 0.8572910024474064,
192
- "rewards/concensus_correctness_reward_func": 0.08016666769981384,
193
- "rewards/consensus_reward_func": 0.0,
194
  "rewards/cumulative_reward_2": 0.0,
195
- "rewards/final_correctness_reward_func": 0.16666666666666666,
196
- "rewards/question_recreation_reward_func": 0.471900115100046,
197
  "rewards/soft_format_reward_func": 0.0,
198
  "rewards/strict_format_reward_func": 0.0,
199
- "rewards/xmlcount_reward_func": 0.3231666736925642,
200
  "step": 20
201
  },
202
  {
203
  "epoch": 5.0,
204
  "step": 20,
205
  "total_flos": 0.0,
206
- "train_loss": 2.26292759180069e-06,
207
- "train_runtime": 300.3933,
208
- "train_samples_per_second": 1.065,
209
- "train_steps_per_second": 0.067
210
  }
211
  ],
212
  "logging_steps": 2,
@@ -227,7 +227,7 @@
227
  }
228
  },
229
  "total_flos": 0.0,
230
- "train_batch_size": 2,
231
  "trial_name": null,
232
  "trial_params": null
233
  }
 
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "completion_length": 370.5,
14
+ "epoch": 0.5333333333333333,
15
+ "grad_norm": 0.6320868134498596,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
  "loss": -0.0,
19
+ "reward": 1.178517110645771,
20
+ "reward_std": 0.9825490638613701,
21
+ "rewards/concensus_correctness_reward_func": 0.04531250335276127,
22
+ "rewards/consensus_reward_func": 0.125,
23
  "rewards/cumulative_reward_2": 0.0,
24
+ "rewards/final_correctness_reward_func": 0.25,
25
+ "rewards/question_recreation_reward_func": 0.420610835775733,
26
  "rewards/soft_format_reward_func": 0.0,
27
+ "rewards/strict_format_reward_func": 0.015625,
28
+ "rewards/xmlcount_reward_func": 0.3219687547534704,
29
  "step": 2
30
  },
31
  {
32
+ "completion_length": 328.92857142857144,
33
  "epoch": 1.0,
34
+ "grad_norm": 0.47581061720848083,
35
+ "kl": 0.0009534425917081535,
36
  "learning_rate": 4.864543104251586e-07,
37
  "loss": 0.0,
38
+ "reward": 1.2720517558710915,
39
+ "reward_std": 1.0473234908921378,
40
+ "rewards/concensus_correctness_reward_func": 0.0,
41
+ "rewards/consensus_reward_func": 0.14285714285714285,
42
  "rewards/cumulative_reward_2": 0.0,
43
+ "rewards/final_correctness_reward_func": 0.42857142857142855,
44
+ "rewards/question_recreation_reward_func": 0.4950517469218799,
45
+ "rewards/soft_format_reward_func": 0.017857142857142856,
46
  "rewards/strict_format_reward_func": 0.0,
47
+ "rewards/xmlcount_reward_func": 0.18771429998534067,
48
  "step": 4
49
  },
50
  {
51
+ "completion_length": 390.96875,
52
+ "epoch": 1.5333333333333332,
53
+ "grad_norm": 0.5493902564048767,
54
+ "kl": 0.0007869710389059037,
55
  "learning_rate": 4.472851273490984e-07,
56
  "loss": 0.0,
57
+ "reward": 1.5517104975879192,
58
+ "reward_std": 1.0216393284499645,
59
+ "rewards/concensus_correctness_reward_func": 0.06012500077486038,
60
+ "rewards/consensus_reward_func": 0.1875,
61
  "rewards/cumulative_reward_2": 0.0,
62
+ "rewards/final_correctness_reward_func": 0.375,
63
+ "rewards/question_recreation_reward_func": 0.4212729502469301,
64
  "rewards/soft_format_reward_func": 0.0,
65
+ "rewards/strict_format_reward_func": 0.03125,
66
+ "rewards/xmlcount_reward_func": 0.4765625037252903,
67
  "step": 6
68
  },
69
  {
70
+ "completion_length": 387.5,
71
  "epoch": 2.0,
72
+ "grad_norm": 0.38475939631462097,
73
+ "kl": 0.0007659006265125104,
74
  "learning_rate": 3.867370395306068e-07,
75
  "loss": 0.0,
76
+ "reward": 0.7905604477439608,
77
+ "reward_std": 0.9036691273961749,
78
+ "rewards/concensus_correctness_reward_func": 0.0,
79
+ "rewards/consensus_reward_func": 0.07142857142857142,
80
  "rewards/cumulative_reward_2": 0.0,
81
+ "rewards/final_correctness_reward_func": 0.21428571428571427,
82
+ "rewards/question_recreation_reward_func": 0.4089176186493465,
83
  "rewards/soft_format_reward_func": 0.0,
84
+ "rewards/strict_format_reward_func": 0.017857142857142856,
85
+ "rewards/xmlcount_reward_func": 0.0780714319220611,
86
  "step": 8
87
  },
88
  {
89
+ "completion_length": 298.375,
90
+ "epoch": 2.533333333333333,
91
+ "grad_norm": 4.775766372680664,
92
+ "kl": 0.0009655183021095581,
93
  "learning_rate": 3.1137137178519977e-07,
94
  "loss": 0.0,
95
+ "reward": 1.925378292798996,
96
+ "reward_std": 1.332388810813427,
97
+ "rewards/concensus_correctness_reward_func": 0.09481250308454037,
98
+ "rewards/consensus_reward_func": 0.3125,
99
  "rewards/cumulative_reward_2": 0.0,
100
+ "rewards/final_correctness_reward_func": 0.5625,
101
+ "rewards/question_recreation_reward_func": 0.524846994318068,
102
  "rewards/soft_format_reward_func": 0.0,
103
+ "rewards/strict_format_reward_func": 0.015625,
104
+ "rewards/xmlcount_reward_func": 0.41509374789893627,
105
  "step": 10
106
  },
107
  {
108
+ "completion_length": 359.85714285714283,
109
  "epoch": 3.0,
110
+ "grad_norm": 0.6231438517570496,
111
+ "kl": 0.0008056558934705598,
112
  "learning_rate": 2.2935516363191693e-07,
113
  "loss": 0.0,
114
+ "reward": 1.1937377282551356,
115
+ "reward_std": 1.1367330593722207,
116
+ "rewards/concensus_correctness_reward_func": 0.0,
117
+ "rewards/consensus_reward_func": 0.2857142857142857,
118
  "rewards/cumulative_reward_2": 0.0,
119
+ "rewards/final_correctness_reward_func": 0.21428571428571427,
120
+ "rewards/question_recreation_reward_func": 0.4232020058802196,
121
  "rewards/soft_format_reward_func": 0.0,
122
+ "rewards/strict_format_reward_func": 0.017857142857142856,
123
+ "rewards/xmlcount_reward_func": 0.25267857419592993,
124
  "step": 12
125
  },
126
  {
127
+ "completion_length": 391.75,
128
+ "epoch": 3.533333333333333,
129
+ "grad_norm": 0.671714723110199,
130
+ "kl": 0.0007653082866454497,
131
  "learning_rate": 1.4957614383675767e-07,
132
  "loss": 0.0,
133
+ "reward": 2.090083334594965,
134
+ "reward_std": 1.2006787694990635,
135
+ "rewards/concensus_correctness_reward_func": 0.427249975502491,
136
+ "rewards/consensus_reward_func": 0.375,
137
  "rewards/cumulative_reward_2": 0.0,
138
+ "rewards/final_correctness_reward_func": 0.4375,
139
+ "rewards/question_recreation_reward_func": 0.4738645385950804,
140
  "rewards/soft_format_reward_func": 0.0,
141
+ "rewards/strict_format_reward_func": 0.0625,
142
+ "rewards/xmlcount_reward_func": 0.3139687478542328,
143
  "step": 14
144
  },
145
  {
146
+ "completion_length": 279.60714285714283,
147
  "epoch": 4.0,
148
+ "grad_norm": 0.6454631090164185,
149
+ "kl": 0.0009609005646780133,
150
  "learning_rate": 8.067960709356478e-08,
151
  "loss": 0.0,
152
+ "reward": 2.0080738365650177,
153
+ "reward_std": 1.9522871311221803,
154
+ "rewards/concensus_correctness_reward_func": 0.7142857142857143,
155
+ "rewards/consensus_reward_func": 0.07142857142857142,
156
  "rewards/cumulative_reward_2": 0.0,
157
+ "rewards/final_correctness_reward_func": 0.2857142857142857,
158
+ "rewards/question_recreation_reward_func": 0.5558595657348633,
159
  "rewards/soft_format_reward_func": 0.0,
160
  "rewards/strict_format_reward_func": 0.0,
161
+ "rewards/xmlcount_reward_func": 0.38078572281769346,
162
  "step": 16
163
  },
164
  {
165
+ "completion_length": 352.21875,
166
+ "epoch": 4.533333333333333,
167
+ "grad_norm": 0.7853459119796753,
168
+ "kl": 0.0008694051721249707,
169
  "learning_rate": 3.013156219837776e-08,
170
  "loss": 0.0,
171
+ "reward": 1.630488459020853,
172
+ "reward_std": 1.1447128355503082,
173
+ "rewards/concensus_correctness_reward_func": 0.07325000315904617,
174
+ "rewards/consensus_reward_func": 0.125,
175
  "rewards/cumulative_reward_2": 0.0,
176
+ "rewards/final_correctness_reward_func": 0.6875,
177
+ "rewards/question_recreation_reward_func": 0.4198634847998619,
178
  "rewards/soft_format_reward_func": 0.0,
179
  "rewards/strict_format_reward_func": 0.0,
180
+ "rewards/xmlcount_reward_func": 0.32487500412389636,
181
  "step": 18
182
  },
183
  {
184
+ "completion_length": 347.42857142857144,
185
  "epoch": 5.0,
186
+ "grad_norm": 0.8366572856903076,
187
+ "kl": 0.0008741397344108139,
188
  "learning_rate": 3.4096741493194193e-09,
189
  "loss": 0.0,
190
+ "reward": 1.2023059981209892,
191
+ "reward_std": 0.7625741745744433,
192
+ "rewards/concensus_correctness_reward_func": -0.018285714089870453,
193
+ "rewards/consensus_reward_func": 0.35714285714285715,
194
  "rewards/cumulative_reward_2": 0.0,
195
+ "rewards/final_correctness_reward_func": 0.0,
196
+ "rewards/question_recreation_reward_func": 0.4690203091927937,
197
  "rewards/soft_format_reward_func": 0.0,
198
  "rewards/strict_format_reward_func": 0.0,
199
+ "rewards/xmlcount_reward_func": 0.39442857993500574,
200
  "step": 20
201
  },
202
  {
203
  "epoch": 5.0,
204
  "step": 20,
205
  "total_flos": 0.0,
206
+ "train_loss": 7.334376395995435e-07,
207
+ "train_runtime": 1213.9576,
208
+ "train_samples_per_second": 0.264,
209
+ "train_steps_per_second": 0.016
210
  }
211
  ],
212
  "logging_steps": 2,
 
227
  }
228
  },
229
  "total_flos": 0.0,
230
+ "train_batch_size": 4,
231
  "trial_name": null,
232
  "trial_params": null
233
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e43aa6fe32facf9e7aaa21f4a7707ea5472944f1653e3413312342178d2b2bd
3
- size 5880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a841150f51cc8ee1957f21d60dd58abd68d7d752067ab1912afcdf9f5c2cab1
3
+ size 5944