shuheng commited on
Commit
950766c
verified
1 Parent(s): 684b875

End of training

Browse files
.gitattributes CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ eval_nbest_predictions.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -4,6 +4,8 @@ license: llama3.2
4
  base_model: meta-llama/Llama-3.2-1B
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: squad_llama_finetuned
9
  results: []
@@ -14,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # squad_llama_finetuned
16
 
17
- This model is a fine-tuned version of [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) on an unknown dataset.
18
 
19
  ## Model description
20
 
 
4
  base_model: meta-llama/Llama-3.2-1B
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - squad
9
  model-index:
10
  - name: squad_llama_finetuned
11
  results: []
 
16
 
17
  # squad_llama_finetuned
18
 
19
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) on the squad dataset.
20
 
21
  ## Model description
22
 
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "eval_exact_match": 19.441816461684013,
4
+ "eval_f1": 29.861733635502148,
5
+ "eval_runtime": 244.9635,
6
+ "eval_samples": 10787,
7
+ "eval_samples_per_second": 44.035,
8
+ "eval_steps_per_second": 5.507,
9
+ "total_flos": 3.971230131335731e+17,
10
+ "train_loss": 3.130299679309049,
11
+ "train_runtime": 13722.7795,
12
+ "train_samples": 88559,
13
+ "train_samples_per_second": 12.907,
14
+ "train_steps_per_second": 1.076
15
+ }
eval_nbest_predictions.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbd692d745a647da9c31dabd52c9e1415964aa602b23d8fc69a917c672a674c0
3
+ size 50840270
eval_predictions.json ADDED
The diff for this file is too large to render. See raw diff
 
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "eval_exact_match": 19.441816461684013,
4
+ "eval_f1": 29.861733635502148,
5
+ "eval_runtime": 244.9635,
6
+ "eval_samples": 10787,
7
+ "eval_samples_per_second": 44.035,
8
+ "eval_steps_per_second": 5.507
9
+ }
runs/Nov15_23-17-58_xgpi8/events.out.tfevents.1731698261.xgpi8.290848.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:779905ce594ceed352860a2b5f51542d074def75adc9c5de577daf99e59e5cb8
3
+ size 412
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "total_flos": 3.971230131335731e+17,
4
+ "train_loss": 3.130299679309049,
5
+ "train_runtime": 13722.7795,
6
+ "train_samples": 88559,
7
+ "train_samples_per_second": 12.907,
8
+ "train_steps_per_second": 1.076
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,245 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 14760,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.06775067750677506,
13
+ "grad_norm": 13.977642059326172,
14
+ "learning_rate": 2.8983739837398373e-05,
15
+ "loss": 4.614,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.13550135501355012,
20
+ "grad_norm": 8.07140064239502,
21
+ "learning_rate": 2.7967479674796748e-05,
22
+ "loss": 4.1059,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.2032520325203252,
27
+ "grad_norm": 13.195443153381348,
28
+ "learning_rate": 2.695121951219512e-05,
29
+ "loss": 3.9346,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 0.27100271002710025,
34
+ "grad_norm": 7.072673797607422,
35
+ "learning_rate": 2.59349593495935e-05,
36
+ "loss": 3.8225,
37
+ "step": 2000
38
+ },
39
+ {
40
+ "epoch": 0.33875338753387535,
41
+ "grad_norm": 8.832470893859863,
42
+ "learning_rate": 2.491869918699187e-05,
43
+ "loss": 3.7412,
44
+ "step": 2500
45
+ },
46
+ {
47
+ "epoch": 0.4065040650406504,
48
+ "grad_norm": 6.735231876373291,
49
+ "learning_rate": 2.3902439024390246e-05,
50
+ "loss": 3.6643,
51
+ "step": 3000
52
+ },
53
+ {
54
+ "epoch": 0.4742547425474255,
55
+ "grad_norm": 7.775442123413086,
56
+ "learning_rate": 2.2886178861788618e-05,
57
+ "loss": 3.6398,
58
+ "step": 3500
59
+ },
60
+ {
61
+ "epoch": 0.5420054200542005,
62
+ "grad_norm": 10.120951652526855,
63
+ "learning_rate": 2.186991869918699e-05,
64
+ "loss": 3.5632,
65
+ "step": 4000
66
+ },
67
+ {
68
+ "epoch": 0.6097560975609756,
69
+ "grad_norm": 6.712512493133545,
70
+ "learning_rate": 2.0853658536585365e-05,
71
+ "loss": 3.5075,
72
+ "step": 4500
73
+ },
74
+ {
75
+ "epoch": 0.6775067750677507,
76
+ "grad_norm": 9.954794883728027,
77
+ "learning_rate": 1.983739837398374e-05,
78
+ "loss": 3.4741,
79
+ "step": 5000
80
+ },
81
+ {
82
+ "epoch": 0.7452574525745257,
83
+ "grad_norm": 8.68674087524414,
84
+ "learning_rate": 1.8821138211382116e-05,
85
+ "loss": 3.4156,
86
+ "step": 5500
87
+ },
88
+ {
89
+ "epoch": 0.8130081300813008,
90
+ "grad_norm": 10.184517860412598,
91
+ "learning_rate": 1.7804878048780488e-05,
92
+ "loss": 3.4252,
93
+ "step": 6000
94
+ },
95
+ {
96
+ "epoch": 0.8807588075880759,
97
+ "grad_norm": 10.719181060791016,
98
+ "learning_rate": 1.6788617886178863e-05,
99
+ "loss": 3.3686,
100
+ "step": 6500
101
+ },
102
+ {
103
+ "epoch": 0.948509485094851,
104
+ "grad_norm": 7.8517560958862305,
105
+ "learning_rate": 1.5772357723577235e-05,
106
+ "loss": 3.3269,
107
+ "step": 7000
108
+ },
109
+ {
110
+ "epoch": 1.016260162601626,
111
+ "grad_norm": 11.628539085388184,
112
+ "learning_rate": 1.475609756097561e-05,
113
+ "loss": 3.208,
114
+ "step": 7500
115
+ },
116
+ {
117
+ "epoch": 1.084010840108401,
118
+ "grad_norm": 20.40143585205078,
119
+ "learning_rate": 1.3739837398373984e-05,
120
+ "loss": 2.7868,
121
+ "step": 8000
122
+ },
123
+ {
124
+ "epoch": 1.151761517615176,
125
+ "grad_norm": 18.49318504333496,
126
+ "learning_rate": 1.2723577235772358e-05,
127
+ "loss": 2.7465,
128
+ "step": 8500
129
+ },
130
+ {
131
+ "epoch": 1.2195121951219512,
132
+ "grad_norm": 18.326629638671875,
133
+ "learning_rate": 1.1707317073170733e-05,
134
+ "loss": 2.6684,
135
+ "step": 9000
136
+ },
137
+ {
138
+ "epoch": 1.2872628726287263,
139
+ "grad_norm": 23.60063362121582,
140
+ "learning_rate": 1.0691056910569105e-05,
141
+ "loss": 2.6501,
142
+ "step": 9500
143
+ },
144
+ {
145
+ "epoch": 1.3550135501355014,
146
+ "grad_norm": 27.22466278076172,
147
+ "learning_rate": 9.674796747967479e-06,
148
+ "loss": 2.656,
149
+ "step": 10000
150
+ },
151
+ {
152
+ "epoch": 1.4227642276422765,
153
+ "grad_norm": 24.0554141998291,
154
+ "learning_rate": 8.658536585365854e-06,
155
+ "loss": 2.5892,
156
+ "step": 10500
157
+ },
158
+ {
159
+ "epoch": 1.4905149051490514,
160
+ "grad_norm": 22.511192321777344,
161
+ "learning_rate": 7.642276422764228e-06,
162
+ "loss": 2.5712,
163
+ "step": 11000
164
+ },
165
+ {
166
+ "epoch": 1.5582655826558267,
167
+ "grad_norm": 25.18529510498047,
168
+ "learning_rate": 6.626016260162602e-06,
169
+ "loss": 2.5993,
170
+ "step": 11500
171
+ },
172
+ {
173
+ "epoch": 1.6260162601626016,
174
+ "grad_norm": 33.19571304321289,
175
+ "learning_rate": 5.609756097560976e-06,
176
+ "loss": 2.5777,
177
+ "step": 12000
178
+ },
179
+ {
180
+ "epoch": 1.6937669376693767,
181
+ "grad_norm": 22.047767639160156,
182
+ "learning_rate": 4.5934959349593495e-06,
183
+ "loss": 2.5205,
184
+ "step": 12500
185
+ },
186
+ {
187
+ "epoch": 1.7615176151761518,
188
+ "grad_norm": 25.974809646606445,
189
+ "learning_rate": 3.5772357723577236e-06,
190
+ "loss": 2.5055,
191
+ "step": 13000
192
+ },
193
+ {
194
+ "epoch": 1.8292682926829267,
195
+ "grad_norm": 27.9240779876709,
196
+ "learning_rate": 2.5609756097560977e-06,
197
+ "loss": 2.4599,
198
+ "step": 13500
199
+ },
200
+ {
201
+ "epoch": 1.897018970189702,
202
+ "grad_norm": 28.034526824951172,
203
+ "learning_rate": 1.5447154471544717e-06,
204
+ "loss": 2.4843,
205
+ "step": 14000
206
+ },
207
+ {
208
+ "epoch": 1.9647696476964769,
209
+ "grad_norm": 28.660612106323242,
210
+ "learning_rate": 5.284552845528455e-07,
211
+ "loss": 2.4865,
212
+ "step": 14500
213
+ },
214
+ {
215
+ "epoch": 2.0,
216
+ "step": 14760,
217
+ "total_flos": 3.971230131335731e+17,
218
+ "train_loss": 3.130299679309049,
219
+ "train_runtime": 13722.7795,
220
+ "train_samples_per_second": 12.907,
221
+ "train_steps_per_second": 1.076
222
+ }
223
+ ],
224
+ "logging_steps": 500,
225
+ "max_steps": 14760,
226
+ "num_input_tokens_seen": 0,
227
+ "num_train_epochs": 2,
228
+ "save_steps": 500,
229
+ "stateful_callbacks": {
230
+ "TrainerControl": {
231
+ "args": {
232
+ "should_epoch_stop": false,
233
+ "should_evaluate": false,
234
+ "should_log": false,
235
+ "should_save": true,
236
+ "should_training_stop": true
237
+ },
238
+ "attributes": {}
239
+ }
240
+ },
241
+ "total_flos": 3.971230131335731e+17,
242
+ "train_batch_size": 12,
243
+ "trial_name": null,
244
+ "trial_params": null
245
+ }