|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 39288, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03817959682345754, |
|
"grad_norm": 47.025718688964844, |
|
"learning_rate": 4.936367338627571e-05, |
|
"loss": 19.8956, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.07635919364691508, |
|
"grad_norm": 39.02047348022461, |
|
"learning_rate": 4.872734677255142e-05, |
|
"loss": 2.0594, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.11453879047037263, |
|
"grad_norm": 82.10893249511719, |
|
"learning_rate": 4.8091020158827126e-05, |
|
"loss": 1.924, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.15271838729383017, |
|
"grad_norm": 25.970558166503906, |
|
"learning_rate": 4.7454693545102835e-05, |
|
"loss": 1.6354, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.19089798411728773, |
|
"grad_norm": 46.208919525146484, |
|
"learning_rate": 4.6818366931378536e-05, |
|
"loss": 1.5016, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.22907758094074526, |
|
"grad_norm": 35.68703842163086, |
|
"learning_rate": 4.6182040317654244e-05, |
|
"loss": 1.5022, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.2672571777642028, |
|
"grad_norm": 65.33502960205078, |
|
"learning_rate": 4.554571370392995e-05, |
|
"loss": 1.423, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.30543677458766033, |
|
"grad_norm": 33.0444450378418, |
|
"learning_rate": 4.490938709020566e-05, |
|
"loss": 1.3971, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.3436163714111179, |
|
"grad_norm": 18.407243728637695, |
|
"learning_rate": 4.427306047648137e-05, |
|
"loss": 1.308, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.38179596823457546, |
|
"grad_norm": 62.20368194580078, |
|
"learning_rate": 4.3636733862757076e-05, |
|
"loss": 1.3365, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.41997556505803296, |
|
"grad_norm": 18.440366744995117, |
|
"learning_rate": 4.3000407249032784e-05, |
|
"loss": 1.1873, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.4581551618814905, |
|
"grad_norm": 21.170751571655273, |
|
"learning_rate": 4.236408063530849e-05, |
|
"loss": 1.212, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.4963347587049481, |
|
"grad_norm": 38.78303146362305, |
|
"learning_rate": 4.17277540215842e-05, |
|
"loss": 1.2951, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.5345143555284056, |
|
"grad_norm": 15.414962768554688, |
|
"learning_rate": 4.109142740785991e-05, |
|
"loss": 1.2291, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.5726939523518632, |
|
"grad_norm": 17.313640594482422, |
|
"learning_rate": 4.0455100794135616e-05, |
|
"loss": 1.1234, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.6108735491753207, |
|
"grad_norm": 24.39589500427246, |
|
"learning_rate": 3.9818774180411324e-05, |
|
"loss": 1.2496, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.6490531459987783, |
|
"grad_norm": 29.411834716796875, |
|
"learning_rate": 3.918244756668703e-05, |
|
"loss": 1.1647, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.6872327428222358, |
|
"grad_norm": 24.564817428588867, |
|
"learning_rate": 3.854612095296274e-05, |
|
"loss": 1.1065, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.7254123396456933, |
|
"grad_norm": 48.59101104736328, |
|
"learning_rate": 3.790979433923845e-05, |
|
"loss": 1.1279, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.7635919364691509, |
|
"grad_norm": 37.30824661254883, |
|
"learning_rate": 3.7273467725514156e-05, |
|
"loss": 1.0621, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.8017715332926084, |
|
"grad_norm": 40.20986557006836, |
|
"learning_rate": 3.6637141111789864e-05, |
|
"loss": 1.1039, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.8399511301160659, |
|
"grad_norm": 7.365723609924316, |
|
"learning_rate": 3.600081449806557e-05, |
|
"loss": 1.0561, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.8781307269395235, |
|
"grad_norm": 10.872908592224121, |
|
"learning_rate": 3.5364487884341273e-05, |
|
"loss": 1.0061, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.916310323762981, |
|
"grad_norm": 37.47850799560547, |
|
"learning_rate": 3.472816127061698e-05, |
|
"loss": 1.1034, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.9544899205864386, |
|
"grad_norm": 59.97223663330078, |
|
"learning_rate": 3.409183465689269e-05, |
|
"loss": 1.058, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.9926695174098962, |
|
"grad_norm": 17.347299575805664, |
|
"learning_rate": 3.34555080431684e-05, |
|
"loss": 0.9714, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.0308491142333538, |
|
"grad_norm": 47.64847183227539, |
|
"learning_rate": 3.2819181429444106e-05, |
|
"loss": 0.9232, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.0690287110568113, |
|
"grad_norm": 23.9332332611084, |
|
"learning_rate": 3.2182854815719814e-05, |
|
"loss": 0.8723, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.1072083078802688, |
|
"grad_norm": 24.243032455444336, |
|
"learning_rate": 3.154652820199552e-05, |
|
"loss": 0.9376, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.1453879047037263, |
|
"grad_norm": 18.394193649291992, |
|
"learning_rate": 3.091020158827123e-05, |
|
"loss": 0.9212, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.1835675015271838, |
|
"grad_norm": 33.30715560913086, |
|
"learning_rate": 3.027387497454694e-05, |
|
"loss": 0.9196, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.2217470983506413, |
|
"grad_norm": 116.27849578857422, |
|
"learning_rate": 2.9637548360822642e-05, |
|
"loss": 0.8325, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.2599266951740988, |
|
"grad_norm": 33.555179595947266, |
|
"learning_rate": 2.900122174709835e-05, |
|
"loss": 0.8746, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.2981062919975566, |
|
"grad_norm": 35.980411529541016, |
|
"learning_rate": 2.836489513337406e-05, |
|
"loss": 0.8561, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.336285888821014, |
|
"grad_norm": 42.04365921020508, |
|
"learning_rate": 2.7728568519649766e-05, |
|
"loss": 0.867, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.3744654856444716, |
|
"grad_norm": 43.86580276489258, |
|
"learning_rate": 2.7092241905925474e-05, |
|
"loss": 0.8331, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.412645082467929, |
|
"grad_norm": 44.562686920166016, |
|
"learning_rate": 2.6455915292201182e-05, |
|
"loss": 0.8354, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.4508246792913866, |
|
"grad_norm": 15.569896697998047, |
|
"learning_rate": 2.581958867847689e-05, |
|
"loss": 0.8514, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.4890042761148443, |
|
"grad_norm": 58.31094741821289, |
|
"learning_rate": 2.51832620647526e-05, |
|
"loss": 0.8712, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.5271838729383018, |
|
"grad_norm": 15.770406723022461, |
|
"learning_rate": 2.4546935451028307e-05, |
|
"loss": 0.8085, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.5653634697617593, |
|
"grad_norm": 29.918487548828125, |
|
"learning_rate": 2.3910608837304015e-05, |
|
"loss": 0.7545, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.6035430665852168, |
|
"grad_norm": 112.88292694091797, |
|
"learning_rate": 2.3274282223579723e-05, |
|
"loss": 0.8347, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.6417226634086743, |
|
"grad_norm": 16.153079986572266, |
|
"learning_rate": 2.2637955609855427e-05, |
|
"loss": 0.8042, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.679902260232132, |
|
"grad_norm": 17.334716796875, |
|
"learning_rate": 2.2001628996131135e-05, |
|
"loss": 0.8368, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.7180818570555894, |
|
"grad_norm": 22.649688720703125, |
|
"learning_rate": 2.1365302382406843e-05, |
|
"loss": 0.7438, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.756261453879047, |
|
"grad_norm": 23.706069946289062, |
|
"learning_rate": 2.072897576868255e-05, |
|
"loss": 0.8163, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.7944410507025046, |
|
"grad_norm": 60.12598419189453, |
|
"learning_rate": 2.0092649154958256e-05, |
|
"loss": 0.8018, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.832620647525962, |
|
"grad_norm": 88.19845581054688, |
|
"learning_rate": 1.9456322541233964e-05, |
|
"loss": 0.7776, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.8708002443494198, |
|
"grad_norm": 18.326126098632812, |
|
"learning_rate": 1.8819995927509672e-05, |
|
"loss": 0.7694, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.9089798411728771, |
|
"grad_norm": 33.15532684326172, |
|
"learning_rate": 1.818366931378538e-05, |
|
"loss": 0.7517, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.9471594379963348, |
|
"grad_norm": 11.88496208190918, |
|
"learning_rate": 1.7547342700061088e-05, |
|
"loss": 0.7802, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.9853390348197923, |
|
"grad_norm": 53.72198486328125, |
|
"learning_rate": 1.6911016086336796e-05, |
|
"loss": 0.7699, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.02351863164325, |
|
"grad_norm": 13.519152641296387, |
|
"learning_rate": 1.6274689472612504e-05, |
|
"loss": 0.7362, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.0616982284667076, |
|
"grad_norm": 21.352752685546875, |
|
"learning_rate": 1.5638362858888212e-05, |
|
"loss": 0.7043, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.099877825290165, |
|
"grad_norm": 39.630210876464844, |
|
"learning_rate": 1.500203624516392e-05, |
|
"loss": 0.6899, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 2.1380574221136226, |
|
"grad_norm": 15.605965614318848, |
|
"learning_rate": 1.4365709631439625e-05, |
|
"loss": 0.7241, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.17623701893708, |
|
"grad_norm": 12.23674201965332, |
|
"learning_rate": 1.3729383017715333e-05, |
|
"loss": 0.6201, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 2.2144166157605376, |
|
"grad_norm": 15.482099533081055, |
|
"learning_rate": 1.309305640399104e-05, |
|
"loss": 0.6966, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.2525962125839953, |
|
"grad_norm": 25.2972469329834, |
|
"learning_rate": 1.2456729790266749e-05, |
|
"loss": 0.5878, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 2.2907758094074526, |
|
"grad_norm": 18.335323333740234, |
|
"learning_rate": 1.1820403176542457e-05, |
|
"loss": 0.6468, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.3289554062309104, |
|
"grad_norm": 44.75697326660156, |
|
"learning_rate": 1.1184076562818163e-05, |
|
"loss": 0.6888, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 2.3671350030543676, |
|
"grad_norm": 9.713996887207031, |
|
"learning_rate": 1.0547749949093871e-05, |
|
"loss": 0.6584, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.4053145998778254, |
|
"grad_norm": 19.04861831665039, |
|
"learning_rate": 9.91142333536958e-06, |
|
"loss": 0.6356, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 2.4434941967012827, |
|
"grad_norm": 16.914947509765625, |
|
"learning_rate": 9.275096721645286e-06, |
|
"loss": 0.6188, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.4816737935247404, |
|
"grad_norm": 25.262630462646484, |
|
"learning_rate": 8.638770107920995e-06, |
|
"loss": 0.6062, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 2.5198533903481977, |
|
"grad_norm": 49.74958419799805, |
|
"learning_rate": 8.002443494196702e-06, |
|
"loss": 0.6554, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.5580329871716554, |
|
"grad_norm": 35.74587631225586, |
|
"learning_rate": 7.36611688047241e-06, |
|
"loss": 0.6523, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 2.596212583995113, |
|
"grad_norm": 22.55206298828125, |
|
"learning_rate": 6.729790266748116e-06, |
|
"loss": 0.6333, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.6343921808185704, |
|
"grad_norm": 19.39708709716797, |
|
"learning_rate": 6.093463653023825e-06, |
|
"loss": 0.6269, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 2.672571777642028, |
|
"grad_norm": 65.50653839111328, |
|
"learning_rate": 5.457137039299532e-06, |
|
"loss": 0.6825, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 2.7107513744654854, |
|
"grad_norm": 10.603687286376953, |
|
"learning_rate": 4.820810425575239e-06, |
|
"loss": 0.6289, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 2.748930971288943, |
|
"grad_norm": 12.743573188781738, |
|
"learning_rate": 4.184483811850947e-06, |
|
"loss": 0.6369, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 2.787110568112401, |
|
"grad_norm": 8.957477569580078, |
|
"learning_rate": 3.5481571981266544e-06, |
|
"loss": 0.6095, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 2.825290164935858, |
|
"grad_norm": 38.0220947265625, |
|
"learning_rate": 2.911830584402362e-06, |
|
"loss": 0.5497, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 2.863469761759316, |
|
"grad_norm": 25.39042854309082, |
|
"learning_rate": 2.27550397067807e-06, |
|
"loss": 0.5781, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 2.901649358582773, |
|
"grad_norm": 11.314273834228516, |
|
"learning_rate": 1.6391773569537775e-06, |
|
"loss": 0.6227, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 2.939828955406231, |
|
"grad_norm": 38.86436080932617, |
|
"learning_rate": 1.0028507432294848e-06, |
|
"loss": 0.5966, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 2.9780085522296886, |
|
"grad_norm": 11.880073547363281, |
|
"learning_rate": 3.6652412950519244e-07, |
|
"loss": 0.5901, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 39288, |
|
"total_flos": 5175762253906176.0, |
|
"train_loss": 1.154250169620813, |
|
"train_runtime": 1582.2568, |
|
"train_samples_per_second": 198.63, |
|
"train_steps_per_second": 24.83 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 39288, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5175762253906176.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|