starchat2-15b-v0.1 / trainer_state.json
li-muyang's picture
Model save
c2540ef verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 100,
"global_step": 2730,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.001098901098901099,
"grad_norm": 2.5938866235339426,
"learning_rate": 7.326007326007327e-08,
"loss": 1.4331,
"step": 1
},
{
"epoch": 0.005494505494505495,
"grad_norm": 2.3305040971210316,
"learning_rate": 3.6630036630036635e-07,
"loss": 1.3861,
"step": 5
},
{
"epoch": 0.01098901098901099,
"grad_norm": 2.3032315871842037,
"learning_rate": 7.326007326007327e-07,
"loss": 1.3892,
"step": 10
},
{
"epoch": 0.016483516483516484,
"grad_norm": 2.0293111765459386,
"learning_rate": 1.098901098901099e-06,
"loss": 1.3496,
"step": 15
},
{
"epoch": 0.02197802197802198,
"grad_norm": 1.4027172304084468,
"learning_rate": 1.4652014652014654e-06,
"loss": 1.2629,
"step": 20
},
{
"epoch": 0.027472527472527472,
"grad_norm": 1.4093182507371345,
"learning_rate": 1.8315018315018316e-06,
"loss": 1.1094,
"step": 25
},
{
"epoch": 0.03296703296703297,
"grad_norm": 0.6639612973057083,
"learning_rate": 2.197802197802198e-06,
"loss": 1.0554,
"step": 30
},
{
"epoch": 0.038461538461538464,
"grad_norm": 0.6143428370522973,
"learning_rate": 2.564102564102564e-06,
"loss": 0.9925,
"step": 35
},
{
"epoch": 0.04395604395604396,
"grad_norm": 0.49724360049126803,
"learning_rate": 2.930402930402931e-06,
"loss": 0.964,
"step": 40
},
{
"epoch": 0.04945054945054945,
"grad_norm": 0.4775849339023127,
"learning_rate": 3.2967032967032968e-06,
"loss": 0.9479,
"step": 45
},
{
"epoch": 0.054945054945054944,
"grad_norm": 0.33493876785224186,
"learning_rate": 3.663003663003663e-06,
"loss": 0.9174,
"step": 50
},
{
"epoch": 0.06043956043956044,
"grad_norm": 0.32489963376779446,
"learning_rate": 4.0293040293040296e-06,
"loss": 0.8859,
"step": 55
},
{
"epoch": 0.06593406593406594,
"grad_norm": 0.492730880028892,
"learning_rate": 4.395604395604396e-06,
"loss": 0.9012,
"step": 60
},
{
"epoch": 0.07142857142857142,
"grad_norm": 0.3109753841977473,
"learning_rate": 4.761904761904762e-06,
"loss": 0.8766,
"step": 65
},
{
"epoch": 0.07692307692307693,
"grad_norm": 0.2867877510996371,
"learning_rate": 5.128205128205128e-06,
"loss": 0.8571,
"step": 70
},
{
"epoch": 0.08241758241758242,
"grad_norm": 0.27490143167772046,
"learning_rate": 5.494505494505495e-06,
"loss": 0.8589,
"step": 75
},
{
"epoch": 0.08791208791208792,
"grad_norm": 0.28545809911171605,
"learning_rate": 5.860805860805862e-06,
"loss": 0.868,
"step": 80
},
{
"epoch": 0.09340659340659341,
"grad_norm": 0.2757573337248957,
"learning_rate": 6.227106227106228e-06,
"loss": 0.8488,
"step": 85
},
{
"epoch": 0.0989010989010989,
"grad_norm": 0.28168284468857463,
"learning_rate": 6.5934065934065935e-06,
"loss": 0.8415,
"step": 90
},
{
"epoch": 0.1043956043956044,
"grad_norm": 0.2789025930294136,
"learning_rate": 6.95970695970696e-06,
"loss": 0.8478,
"step": 95
},
{
"epoch": 0.10989010989010989,
"grad_norm": 0.2793158898424537,
"learning_rate": 7.326007326007326e-06,
"loss": 0.8402,
"step": 100
},
{
"epoch": 0.10989010989010989,
"eval_loss": 0.8307385444641113,
"eval_runtime": 58.2285,
"eval_samples_per_second": 23.665,
"eval_steps_per_second": 0.103,
"step": 100
},
{
"epoch": 0.11538461538461539,
"grad_norm": 0.28714132101165046,
"learning_rate": 7.692307692307694e-06,
"loss": 0.8132,
"step": 105
},
{
"epoch": 0.12087912087912088,
"grad_norm": 0.25219489491882496,
"learning_rate": 8.058608058608059e-06,
"loss": 0.8212,
"step": 110
},
{
"epoch": 0.12637362637362637,
"grad_norm": 0.2621955241800005,
"learning_rate": 8.424908424908426e-06,
"loss": 0.8152,
"step": 115
},
{
"epoch": 0.13186813186813187,
"grad_norm": 0.2699161655817323,
"learning_rate": 8.791208791208792e-06,
"loss": 0.8227,
"step": 120
},
{
"epoch": 0.13736263736263737,
"grad_norm": 0.283439689013222,
"learning_rate": 9.157509157509158e-06,
"loss": 0.7977,
"step": 125
},
{
"epoch": 0.14285714285714285,
"grad_norm": 0.2513858523994278,
"learning_rate": 9.523809523809525e-06,
"loss": 0.7834,
"step": 130
},
{
"epoch": 0.14835164835164835,
"grad_norm": 0.2563724649349777,
"learning_rate": 9.890109890109892e-06,
"loss": 0.8214,
"step": 135
},
{
"epoch": 0.15384615384615385,
"grad_norm": 0.26746604704320437,
"learning_rate": 1.0256410256410256e-05,
"loss": 0.7852,
"step": 140
},
{
"epoch": 0.15934065934065933,
"grad_norm": 0.280443649102156,
"learning_rate": 1.0622710622710623e-05,
"loss": 0.7943,
"step": 145
},
{
"epoch": 0.16483516483516483,
"grad_norm": 0.2661703492350103,
"learning_rate": 1.098901098901099e-05,
"loss": 0.7846,
"step": 150
},
{
"epoch": 0.17032967032967034,
"grad_norm": 0.25295008227570925,
"learning_rate": 1.1355311355311356e-05,
"loss": 0.79,
"step": 155
},
{
"epoch": 0.17582417582417584,
"grad_norm": 0.2572179650619883,
"learning_rate": 1.1721611721611723e-05,
"loss": 0.7945,
"step": 160
},
{
"epoch": 0.1813186813186813,
"grad_norm": 0.28163509900788125,
"learning_rate": 1.2087912087912089e-05,
"loss": 0.7699,
"step": 165
},
{
"epoch": 0.18681318681318682,
"grad_norm": 0.2998101858443158,
"learning_rate": 1.2454212454212456e-05,
"loss": 0.789,
"step": 170
},
{
"epoch": 0.19230769230769232,
"grad_norm": 0.259731613700112,
"learning_rate": 1.2820512820512823e-05,
"loss": 0.7518,
"step": 175
},
{
"epoch": 0.1978021978021978,
"grad_norm": 0.26085689374421417,
"learning_rate": 1.3186813186813187e-05,
"loss": 0.7778,
"step": 180
},
{
"epoch": 0.2032967032967033,
"grad_norm": 0.25933950442397596,
"learning_rate": 1.3553113553113554e-05,
"loss": 0.7752,
"step": 185
},
{
"epoch": 0.2087912087912088,
"grad_norm": 0.2655972864098073,
"learning_rate": 1.391941391941392e-05,
"loss": 0.7754,
"step": 190
},
{
"epoch": 0.21428571428571427,
"grad_norm": 0.3027958618309835,
"learning_rate": 1.4285714285714287e-05,
"loss": 0.7738,
"step": 195
},
{
"epoch": 0.21978021978021978,
"grad_norm": 0.2636680908875259,
"learning_rate": 1.4652014652014653e-05,
"loss": 0.7611,
"step": 200
},
{
"epoch": 0.21978021978021978,
"eval_loss": 0.7792791724205017,
"eval_runtime": 60.1437,
"eval_samples_per_second": 22.912,
"eval_steps_per_second": 0.1,
"step": 200
},
{
"epoch": 0.22527472527472528,
"grad_norm": 0.28418095535412446,
"learning_rate": 1.501831501831502e-05,
"loss": 0.7773,
"step": 205
},
{
"epoch": 0.23076923076923078,
"grad_norm": 0.253579945341443,
"learning_rate": 1.5384615384615387e-05,
"loss": 0.7466,
"step": 210
},
{
"epoch": 0.23626373626373626,
"grad_norm": 0.2709322711520115,
"learning_rate": 1.575091575091575e-05,
"loss": 0.7486,
"step": 215
},
{
"epoch": 0.24175824175824176,
"grad_norm": 0.25784522395547,
"learning_rate": 1.6117216117216118e-05,
"loss": 0.7602,
"step": 220
},
{
"epoch": 0.24725274725274726,
"grad_norm": 0.25800289350422184,
"learning_rate": 1.6483516483516486e-05,
"loss": 0.7593,
"step": 225
},
{
"epoch": 0.25274725274725274,
"grad_norm": 0.4335652422653119,
"learning_rate": 1.6849816849816853e-05,
"loss": 0.7357,
"step": 230
},
{
"epoch": 0.25824175824175827,
"grad_norm": 0.3289531244189634,
"learning_rate": 1.721611721611722e-05,
"loss": 0.7456,
"step": 235
},
{
"epoch": 0.26373626373626374,
"grad_norm": 1.2095875802009959,
"learning_rate": 1.7582417582417584e-05,
"loss": 0.7226,
"step": 240
},
{
"epoch": 0.2692307692307692,
"grad_norm": 0.2550501846115449,
"learning_rate": 1.794871794871795e-05,
"loss": 0.7304,
"step": 245
},
{
"epoch": 0.27472527472527475,
"grad_norm": 0.32241036727466216,
"learning_rate": 1.8315018315018315e-05,
"loss": 0.7322,
"step": 250
},
{
"epoch": 0.2802197802197802,
"grad_norm": 0.23942812421802775,
"learning_rate": 1.8681318681318682e-05,
"loss": 0.7261,
"step": 255
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.2532890631657975,
"learning_rate": 1.904761904761905e-05,
"loss": 0.752,
"step": 260
},
{
"epoch": 0.29120879120879123,
"grad_norm": 0.24364925383721642,
"learning_rate": 1.9413919413919417e-05,
"loss": 0.755,
"step": 265
},
{
"epoch": 0.2967032967032967,
"grad_norm": 0.24181710554401747,
"learning_rate": 1.9780219780219784e-05,
"loss": 0.7554,
"step": 270
},
{
"epoch": 0.3021978021978022,
"grad_norm": 0.23828479627116422,
"learning_rate": 1.9999967302150437e-05,
"loss": 0.7246,
"step": 275
},
{
"epoch": 0.3076923076923077,
"grad_norm": 0.26360754943139847,
"learning_rate": 1.9999599453798523e-05,
"loss": 0.7293,
"step": 280
},
{
"epoch": 0.3131868131868132,
"grad_norm": 0.2624620835053123,
"learning_rate": 1.9998822899867633e-05,
"loss": 0.7282,
"step": 285
},
{
"epoch": 0.31868131868131866,
"grad_norm": 0.24888154268515242,
"learning_rate": 1.9997637672097222e-05,
"loss": 0.7343,
"step": 290
},
{
"epoch": 0.3241758241758242,
"grad_norm": 0.23324534028548077,
"learning_rate": 1.9996043818930153e-05,
"loss": 0.7158,
"step": 295
},
{
"epoch": 0.32967032967032966,
"grad_norm": 0.23821533213282017,
"learning_rate": 1.9994041405510705e-05,
"loss": 0.7361,
"step": 300
},
{
"epoch": 0.32967032967032966,
"eval_loss": 0.7524750828742981,
"eval_runtime": 58.5251,
"eval_samples_per_second": 23.545,
"eval_steps_per_second": 0.103,
"step": 300
},
{
"epoch": 0.33516483516483514,
"grad_norm": 0.2419083485626964,
"learning_rate": 1.999163051368191e-05,
"loss": 0.7392,
"step": 305
},
{
"epoch": 0.34065934065934067,
"grad_norm": 0.23969743620339276,
"learning_rate": 1.9988811241982206e-05,
"loss": 0.722,
"step": 310
},
{
"epoch": 0.34615384615384615,
"grad_norm": 0.24732719030591468,
"learning_rate": 1.9985583705641418e-05,
"loss": 0.7375,
"step": 315
},
{
"epoch": 0.3516483516483517,
"grad_norm": 0.2524078352421696,
"learning_rate": 1.9981948036576045e-05,
"loss": 0.7174,
"step": 320
},
{
"epoch": 0.35714285714285715,
"grad_norm": 0.2450128000419284,
"learning_rate": 1.997790438338385e-05,
"loss": 0.7193,
"step": 325
},
{
"epoch": 0.3626373626373626,
"grad_norm": 0.22547733974392728,
"learning_rate": 1.997345291133783e-05,
"loss": 0.7121,
"step": 330
},
{
"epoch": 0.36813186813186816,
"grad_norm": 0.23499045457962084,
"learning_rate": 1.9968593802379405e-05,
"loss": 0.7067,
"step": 335
},
{
"epoch": 0.37362637362637363,
"grad_norm": 0.24009998902460783,
"learning_rate": 1.9963327255111033e-05,
"loss": 0.724,
"step": 340
},
{
"epoch": 0.3791208791208791,
"grad_norm": 0.27381888592320386,
"learning_rate": 1.9957653484788054e-05,
"loss": 0.7088,
"step": 345
},
{
"epoch": 0.38461538461538464,
"grad_norm": 0.23470929489474682,
"learning_rate": 1.9951572723309918e-05,
"loss": 0.7123,
"step": 350
},
{
"epoch": 0.3901098901098901,
"grad_norm": 0.234461070541495,
"learning_rate": 1.99450852192107e-05,
"loss": 0.7132,
"step": 355
},
{
"epoch": 0.3956043956043956,
"grad_norm": 0.25459804884418324,
"learning_rate": 1.9938191237648924e-05,
"loss": 0.7076,
"step": 360
},
{
"epoch": 0.4010989010989011,
"grad_norm": 0.245152397924957,
"learning_rate": 1.9930891060396757e-05,
"loss": 0.6918,
"step": 365
},
{
"epoch": 0.4065934065934066,
"grad_norm": 0.23860847532916626,
"learning_rate": 1.992318498582846e-05,
"loss": 0.7304,
"step": 370
},
{
"epoch": 0.41208791208791207,
"grad_norm": 0.2519260957646762,
"learning_rate": 1.9915073328908217e-05,
"loss": 0.709,
"step": 375
},
{
"epoch": 0.4175824175824176,
"grad_norm": 0.23360487704945615,
"learning_rate": 1.9906556421177256e-05,
"loss": 0.7258,
"step": 380
},
{
"epoch": 0.4230769230769231,
"grad_norm": 0.23254394099980408,
"learning_rate": 1.989763461074029e-05,
"loss": 0.707,
"step": 385
},
{
"epoch": 0.42857142857142855,
"grad_norm": 0.22834651014549015,
"learning_rate": 1.9888308262251286e-05,
"loss": 0.7085,
"step": 390
},
{
"epoch": 0.4340659340659341,
"grad_norm": 0.22350055020087028,
"learning_rate": 1.987857775689859e-05,
"loss": 0.7153,
"step": 395
},
{
"epoch": 0.43956043956043955,
"grad_norm": 0.2559347137216262,
"learning_rate": 1.9868443492389307e-05,
"loss": 0.6854,
"step": 400
},
{
"epoch": 0.43956043956043955,
"eval_loss": 0.7337242960929871,
"eval_runtime": 57.135,
"eval_samples_per_second": 24.118,
"eval_steps_per_second": 0.105,
"step": 400
},
{
"epoch": 0.44505494505494503,
"grad_norm": 0.27398370193799876,
"learning_rate": 1.985790588293308e-05,
"loss": 0.716,
"step": 405
},
{
"epoch": 0.45054945054945056,
"grad_norm": 0.49293536001562244,
"learning_rate": 1.9846965359225127e-05,
"loss": 0.7187,
"step": 410
},
{
"epoch": 0.45604395604395603,
"grad_norm": 0.23991864498334772,
"learning_rate": 1.9835622368428673e-05,
"loss": 0.6915,
"step": 415
},
{
"epoch": 0.46153846153846156,
"grad_norm": 0.2223500820145743,
"learning_rate": 1.9823877374156647e-05,
"loss": 0.7076,
"step": 420
},
{
"epoch": 0.46703296703296704,
"grad_norm": 0.2327646676318351,
"learning_rate": 1.9811730856452754e-05,
"loss": 0.6865,
"step": 425
},
{
"epoch": 0.4725274725274725,
"grad_norm": 0.23366623666692093,
"learning_rate": 1.9799183311771823e-05,
"loss": 0.697,
"step": 430
},
{
"epoch": 0.47802197802197804,
"grad_norm": 0.23211216480431496,
"learning_rate": 1.9786235252959555e-05,
"loss": 0.6949,
"step": 435
},
{
"epoch": 0.4835164835164835,
"grad_norm": 0.2335049326202,
"learning_rate": 1.977288720923153e-05,
"loss": 0.688,
"step": 440
},
{
"epoch": 0.489010989010989,
"grad_norm": 0.23612214303202572,
"learning_rate": 1.9759139726151597e-05,
"loss": 0.7006,
"step": 445
},
{
"epoch": 0.4945054945054945,
"grad_norm": 0.2296371958742712,
"learning_rate": 1.9744993365609563e-05,
"loss": 0.7096,
"step": 450
},
{
"epoch": 0.5,
"grad_norm": 0.23792816575947545,
"learning_rate": 1.973044870579824e-05,
"loss": 0.6799,
"step": 455
},
{
"epoch": 0.5054945054945055,
"grad_norm": 0.23888207855755858,
"learning_rate": 1.9715506341189795e-05,
"loss": 0.7189,
"step": 460
},
{
"epoch": 0.510989010989011,
"grad_norm": 0.22873637589635504,
"learning_rate": 1.970016688251147e-05,
"loss": 0.7041,
"step": 465
},
{
"epoch": 0.5164835164835165,
"grad_norm": 0.22109885665030304,
"learning_rate": 1.9684430956720613e-05,
"loss": 0.6889,
"step": 470
},
{
"epoch": 0.521978021978022,
"grad_norm": 0.22943401975482253,
"learning_rate": 1.966829920697905e-05,
"loss": 0.701,
"step": 475
},
{
"epoch": 0.5274725274725275,
"grad_norm": 0.24368759924523126,
"learning_rate": 1.9651772292626804e-05,
"loss": 0.6798,
"step": 480
},
{
"epoch": 0.532967032967033,
"grad_norm": 0.22502323506511537,
"learning_rate": 1.963485088915514e-05,
"loss": 0.6732,
"step": 485
},
{
"epoch": 0.5384615384615384,
"grad_norm": 0.23406619299034903,
"learning_rate": 1.961753568817896e-05,
"loss": 0.6808,
"step": 490
},
{
"epoch": 0.5439560439560439,
"grad_norm": 0.24379416412543173,
"learning_rate": 1.959982739740854e-05,
"loss": 0.6732,
"step": 495
},
{
"epoch": 0.5494505494505495,
"grad_norm": 0.22395929004707257,
"learning_rate": 1.9581726740620585e-05,
"loss": 0.6926,
"step": 500
},
{
"epoch": 0.5494505494505495,
"eval_loss": 0.7197047472000122,
"eval_runtime": 57.3825,
"eval_samples_per_second": 24.014,
"eval_steps_per_second": 0.105,
"step": 500
},
{
"epoch": 0.554945054945055,
"grad_norm": 0.23635721578380114,
"learning_rate": 1.9563234457628678e-05,
"loss": 0.7019,
"step": 505
},
{
"epoch": 0.5604395604395604,
"grad_norm": 0.22361881412683712,
"learning_rate": 1.954435130425301e-05,
"loss": 0.6725,
"step": 510
},
{
"epoch": 0.5659340659340659,
"grad_norm": 0.27988489835986297,
"learning_rate": 1.952507805228951e-05,
"loss": 0.6806,
"step": 515
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.22714617789786667,
"learning_rate": 1.9505415489478293e-05,
"loss": 0.6642,
"step": 520
},
{
"epoch": 0.5769230769230769,
"grad_norm": 0.234738186404094,
"learning_rate": 1.9485364419471454e-05,
"loss": 0.6788,
"step": 525
},
{
"epoch": 0.5824175824175825,
"grad_norm": 0.22212354431865575,
"learning_rate": 1.9464925661800247e-05,
"loss": 0.6714,
"step": 530
},
{
"epoch": 0.5879120879120879,
"grad_norm": 0.2260832469755207,
"learning_rate": 1.9444100051841556e-05,
"loss": 0.6742,
"step": 535
},
{
"epoch": 0.5934065934065934,
"grad_norm": 0.22847007927311688,
"learning_rate": 1.9422888440783773e-05,
"loss": 0.6825,
"step": 540
},
{
"epoch": 0.5989010989010989,
"grad_norm": 0.24175296826977385,
"learning_rate": 1.9401291695592e-05,
"loss": 0.6748,
"step": 545
},
{
"epoch": 0.6043956043956044,
"grad_norm": 0.22595935576781812,
"learning_rate": 1.9379310698972618e-05,
"loss": 0.6922,
"step": 550
},
{
"epoch": 0.6098901098901099,
"grad_norm": 0.2357931340905214,
"learning_rate": 1.935694634933721e-05,
"loss": 0.6961,
"step": 555
},
{
"epoch": 0.6153846153846154,
"grad_norm": 0.23436108471733894,
"learning_rate": 1.933419956076584e-05,
"loss": 0.6539,
"step": 560
},
{
"epoch": 0.6208791208791209,
"grad_norm": 0.21960475435288737,
"learning_rate": 1.9311071262969675e-05,
"loss": 0.6815,
"step": 565
},
{
"epoch": 0.6263736263736264,
"grad_norm": 0.22404812837767762,
"learning_rate": 1.9287562401253023e-05,
"loss": 0.6654,
"step": 570
},
{
"epoch": 0.6318681318681318,
"grad_norm": 0.2283357687835303,
"learning_rate": 1.9263673936474662e-05,
"loss": 0.6791,
"step": 575
},
{
"epoch": 0.6373626373626373,
"grad_norm": 0.22007214714544407,
"learning_rate": 1.9239406845008583e-05,
"loss": 0.6707,
"step": 580
},
{
"epoch": 0.6428571428571429,
"grad_norm": 0.21680436366260772,
"learning_rate": 1.921476211870408e-05,
"loss": 0.6875,
"step": 585
},
{
"epoch": 0.6483516483516484,
"grad_norm": 0.2206990118778276,
"learning_rate": 1.918974076484521e-05,
"loss": 0.689,
"step": 590
},
{
"epoch": 0.6538461538461539,
"grad_norm": 0.23773830363891224,
"learning_rate": 1.916434380610963e-05,
"loss": 0.6904,
"step": 595
},
{
"epoch": 0.6593406593406593,
"grad_norm": 0.22280953580435744,
"learning_rate": 1.9138572280526795e-05,
"loss": 0.7125,
"step": 600
},
{
"epoch": 0.6593406593406593,
"eval_loss": 0.709740936756134,
"eval_runtime": 56.8842,
"eval_samples_per_second": 24.225,
"eval_steps_per_second": 0.105,
"step": 600
},
{
"epoch": 0.6648351648351648,
"grad_norm": 0.21474109145946474,
"learning_rate": 1.911242724143552e-05,
"loss": 0.6883,
"step": 605
},
{
"epoch": 0.6703296703296703,
"grad_norm": 0.22132757673296177,
"learning_rate": 1.908590975744094e-05,
"loss": 0.6664,
"step": 610
},
{
"epoch": 0.6758241758241759,
"grad_norm": 0.22862364892697207,
"learning_rate": 1.9059020912370836e-05,
"loss": 0.6628,
"step": 615
},
{
"epoch": 0.6813186813186813,
"grad_norm": 0.22871592729546456,
"learning_rate": 1.9031761805231322e-05,
"loss": 0.6672,
"step": 620
},
{
"epoch": 0.6868131868131868,
"grad_norm": 0.22248256090434504,
"learning_rate": 1.9004133550161953e-05,
"loss": 0.6601,
"step": 625
},
{
"epoch": 0.6923076923076923,
"grad_norm": 0.22365938486859313,
"learning_rate": 1.8976137276390145e-05,
"loss": 0.6826,
"step": 630
},
{
"epoch": 0.6978021978021978,
"grad_norm": 0.2199011359745548,
"learning_rate": 1.894777412818506e-05,
"loss": 0.6641,
"step": 635
},
{
"epoch": 0.7032967032967034,
"grad_norm": 0.22510686451837608,
"learning_rate": 1.891904526481083e-05,
"loss": 0.6859,
"step": 640
},
{
"epoch": 0.7087912087912088,
"grad_norm": 0.21671102889261787,
"learning_rate": 1.8889951860479165e-05,
"loss": 0.6714,
"step": 645
},
{
"epoch": 0.7142857142857143,
"grad_norm": 0.2281707606944894,
"learning_rate": 1.8860495104301346e-05,
"loss": 0.6606,
"step": 650
},
{
"epoch": 0.7197802197802198,
"grad_norm": 0.2307798038244349,
"learning_rate": 1.8830676200239666e-05,
"loss": 0.6743,
"step": 655
},
{
"epoch": 0.7252747252747253,
"grad_norm": 0.2231346316824852,
"learning_rate": 1.8800496367058187e-05,
"loss": 0.6779,
"step": 660
},
{
"epoch": 0.7307692307692307,
"grad_norm": 0.2163289792210269,
"learning_rate": 1.8769956838272937e-05,
"loss": 0.6589,
"step": 665
},
{
"epoch": 0.7362637362637363,
"grad_norm": 0.21367697388384943,
"learning_rate": 1.8739058862101487e-05,
"loss": 0.6663,
"step": 670
},
{
"epoch": 0.7417582417582418,
"grad_norm": 0.2129068920296721,
"learning_rate": 1.8707803701411946e-05,
"loss": 0.6476,
"step": 675
},
{
"epoch": 0.7472527472527473,
"grad_norm": 0.225469317489798,
"learning_rate": 1.8676192633671342e-05,
"loss": 0.6617,
"step": 680
},
{
"epoch": 0.7527472527472527,
"grad_norm": 0.22021116546638458,
"learning_rate": 1.8644226950893394e-05,
"loss": 0.6709,
"step": 685
},
{
"epoch": 0.7582417582417582,
"grad_norm": 0.21699187730804526,
"learning_rate": 1.861190795958573e-05,
"loss": 0.6758,
"step": 690
},
{
"epoch": 0.7637362637362637,
"grad_norm": 0.22500716309661084,
"learning_rate": 1.857923698069646e-05,
"loss": 0.6623,
"step": 695
},
{
"epoch": 0.7692307692307693,
"grad_norm": 0.22095635026785526,
"learning_rate": 1.8546215349560204e-05,
"loss": 0.6662,
"step": 700
},
{
"epoch": 0.7692307692307693,
"eval_loss": 0.7014729976654053,
"eval_runtime": 58.7488,
"eval_samples_per_second": 23.456,
"eval_steps_per_second": 0.102,
"step": 700
},
{
"epoch": 0.7747252747252747,
"grad_norm": 0.2309523218504365,
"learning_rate": 1.8512844415843514e-05,
"loss": 0.6638,
"step": 705
},
{
"epoch": 0.7802197802197802,
"grad_norm": 0.2603467352522321,
"learning_rate": 1.8479125543489694e-05,
"loss": 0.6585,
"step": 710
},
{
"epoch": 0.7857142857142857,
"grad_norm": 0.25516737652528354,
"learning_rate": 1.844506011066308e-05,
"loss": 0.6731,
"step": 715
},
{
"epoch": 0.7912087912087912,
"grad_norm": 0.20905095426392303,
"learning_rate": 1.841064950969268e-05,
"loss": 0.6516,
"step": 720
},
{
"epoch": 0.7967032967032966,
"grad_norm": 0.21250771141180422,
"learning_rate": 1.8375895147015285e-05,
"loss": 0.6597,
"step": 725
},
{
"epoch": 0.8021978021978022,
"grad_norm": 0.21331846730755302,
"learning_rate": 1.8340798443117992e-05,
"loss": 0.6551,
"step": 730
},
{
"epoch": 0.8076923076923077,
"grad_norm": 0.21452172763665409,
"learning_rate": 1.8305360832480118e-05,
"loss": 0.6284,
"step": 735
},
{
"epoch": 0.8131868131868132,
"grad_norm": 0.22124518551235928,
"learning_rate": 1.8269583763514603e-05,
"loss": 0.6629,
"step": 740
},
{
"epoch": 0.8186813186813187,
"grad_norm": 0.20994044274257415,
"learning_rate": 1.8233468698508786e-05,
"loss": 0.6379,
"step": 745
},
{
"epoch": 0.8241758241758241,
"grad_norm": 0.21519633358820012,
"learning_rate": 1.819701711356464e-05,
"loss": 0.657,
"step": 750
},
{
"epoch": 0.8296703296703297,
"grad_norm": 0.2131452043844892,
"learning_rate": 1.8160230498538464e-05,
"loss": 0.6607,
"step": 755
},
{
"epoch": 0.8351648351648352,
"grad_norm": 0.21693178791987028,
"learning_rate": 1.8123110356979955e-05,
"loss": 0.6691,
"step": 760
},
{
"epoch": 0.8406593406593407,
"grad_norm": 0.2122186850793786,
"learning_rate": 1.808565820607078e-05,
"loss": 0.6605,
"step": 765
},
{
"epoch": 0.8461538461538461,
"grad_norm": 0.21668819567741526,
"learning_rate": 1.8047875576562556e-05,
"loss": 0.6518,
"step": 770
},
{
"epoch": 0.8516483516483516,
"grad_norm": 0.21597653641956766,
"learning_rate": 1.8009764012714283e-05,
"loss": 0.6552,
"step": 775
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.21403410821941848,
"learning_rate": 1.7971325072229227e-05,
"loss": 0.6524,
"step": 780
},
{
"epoch": 0.8626373626373627,
"grad_norm": 0.21858357009497015,
"learning_rate": 1.7932560326191265e-05,
"loss": 0.6432,
"step": 785
},
{
"epoch": 0.8681318681318682,
"grad_norm": 0.20835225444574582,
"learning_rate": 1.789347135900066e-05,
"loss": 0.6521,
"step": 790
},
{
"epoch": 0.8736263736263736,
"grad_norm": 0.2147631378246422,
"learning_rate": 1.7854059768309292e-05,
"loss": 0.6635,
"step": 795
},
{
"epoch": 0.8791208791208791,
"grad_norm": 0.21742833943392462,
"learning_rate": 1.7814327164955388e-05,
"loss": 0.6517,
"step": 800
},
{
"epoch": 0.8791208791208791,
"eval_loss": 0.6937416791915894,
"eval_runtime": 58.6272,
"eval_samples_per_second": 23.504,
"eval_steps_per_second": 0.102,
"step": 800
},
{
"epoch": 0.8846153846153846,
"grad_norm": 0.21815615754521292,
"learning_rate": 1.777427517289766e-05,
"loss": 0.6513,
"step": 805
},
{
"epoch": 0.8901098901098901,
"grad_norm": 0.21132299920641998,
"learning_rate": 1.773390542914894e-05,
"loss": 0.6451,
"step": 810
},
{
"epoch": 0.8956043956043956,
"grad_norm": 0.2153981254584012,
"learning_rate": 1.7693219583709266e-05,
"loss": 0.6472,
"step": 815
},
{
"epoch": 0.9010989010989011,
"grad_norm": 0.21516183847901768,
"learning_rate": 1.765221929949845e-05,
"loss": 0.6455,
"step": 820
},
{
"epoch": 0.9065934065934066,
"grad_norm": 0.21374150031251749,
"learning_rate": 1.7610906252288097e-05,
"loss": 0.6599,
"step": 825
},
{
"epoch": 0.9120879120879121,
"grad_norm": 0.22521849898029836,
"learning_rate": 1.7569282130633137e-05,
"loss": 0.6618,
"step": 830
},
{
"epoch": 0.9175824175824175,
"grad_norm": 0.20871378565296117,
"learning_rate": 1.752734863580278e-05,
"loss": 0.6435,
"step": 835
},
{
"epoch": 0.9230769230769231,
"grad_norm": 0.20579905659289663,
"learning_rate": 1.7485107481711014e-05,
"loss": 0.6384,
"step": 840
},
{
"epoch": 0.9285714285714286,
"grad_norm": 0.24382815508466385,
"learning_rate": 1.7442560394846518e-05,
"loss": 0.6433,
"step": 845
},
{
"epoch": 0.9340659340659341,
"grad_norm": 0.20272371923265584,
"learning_rate": 1.739970911420213e-05,
"loss": 0.6647,
"step": 850
},
{
"epoch": 0.9395604395604396,
"grad_norm": 0.21061337303398175,
"learning_rate": 1.7356555391203745e-05,
"loss": 0.6627,
"step": 855
},
{
"epoch": 0.945054945054945,
"grad_norm": 0.22526349877300406,
"learning_rate": 1.7313100989638745e-05,
"loss": 0.6528,
"step": 860
},
{
"epoch": 0.9505494505494505,
"grad_norm": 0.2405922721275717,
"learning_rate": 1.7269347685583913e-05,
"loss": 0.641,
"step": 865
},
{
"epoch": 0.9560439560439561,
"grad_norm": 0.22058952892730063,
"learning_rate": 1.7225297267332815e-05,
"loss": 0.6632,
"step": 870
},
{
"epoch": 0.9615384615384616,
"grad_norm": 0.21098073518114888,
"learning_rate": 1.7180951535322742e-05,
"loss": 0.651,
"step": 875
},
{
"epoch": 0.967032967032967,
"grad_norm": 0.21900984210996413,
"learning_rate": 1.7136312302061097e-05,
"loss": 0.6651,
"step": 880
},
{
"epoch": 0.9725274725274725,
"grad_norm": 0.20847336917155376,
"learning_rate": 1.7091381392051333e-05,
"loss": 0.6364,
"step": 885
},
{
"epoch": 0.978021978021978,
"grad_norm": 0.22051072508635897,
"learning_rate": 1.704616064171836e-05,
"loss": 0.6577,
"step": 890
},
{
"epoch": 0.9835164835164835,
"grad_norm": 0.21269562851079712,
"learning_rate": 1.7000651899333512e-05,
"loss": 0.6457,
"step": 895
},
{
"epoch": 0.989010989010989,
"grad_norm": 0.20529825370314203,
"learning_rate": 1.6954857024938976e-05,
"loss": 0.6234,
"step": 900
},
{
"epoch": 0.989010989010989,
"eval_loss": 0.6869306564331055,
"eval_runtime": 57.713,
"eval_samples_per_second": 23.877,
"eval_steps_per_second": 0.104,
"step": 900
},
{
"epoch": 0.9945054945054945,
"grad_norm": 0.2434572328244998,
"learning_rate": 1.6908777890271794e-05,
"loss": 0.6504,
"step": 905
},
{
"epoch": 1.0,
"grad_norm": 0.21741173594960767,
"learning_rate": 1.686241637868734e-05,
"loss": 0.6335,
"step": 910
},
{
"epoch": 1.0054945054945055,
"grad_norm": 0.2130051055875561,
"learning_rate": 1.6815774385082355e-05,
"loss": 0.5875,
"step": 915
},
{
"epoch": 1.010989010989011,
"grad_norm": 0.21037185236973288,
"learning_rate": 1.6768853815817506e-05,
"loss": 0.5925,
"step": 920
},
{
"epoch": 1.0164835164835164,
"grad_norm": 0.22203340887583678,
"learning_rate": 1.6721656588639444e-05,
"loss": 0.6027,
"step": 925
},
{
"epoch": 1.021978021978022,
"grad_norm": 0.2835076660446732,
"learning_rate": 1.6674184632602447e-05,
"loss": 0.5949,
"step": 930
},
{
"epoch": 1.0274725274725274,
"grad_norm": 0.21059794691991446,
"learning_rate": 1.6626439887989552e-05,
"loss": 0.5952,
"step": 935
},
{
"epoch": 1.032967032967033,
"grad_norm": 0.21091351684344237,
"learning_rate": 1.6578424306233282e-05,
"loss": 0.5912,
"step": 940
},
{
"epoch": 1.0384615384615385,
"grad_norm": 0.21087167061209808,
"learning_rate": 1.653013984983585e-05,
"loss": 0.5862,
"step": 945
},
{
"epoch": 1.043956043956044,
"grad_norm": 0.21380595384587936,
"learning_rate": 1.6481588492288985e-05,
"loss": 0.599,
"step": 950
},
{
"epoch": 1.0494505494505495,
"grad_norm": 0.20714451035024295,
"learning_rate": 1.643277221799323e-05,
"loss": 0.5841,
"step": 955
},
{
"epoch": 1.054945054945055,
"grad_norm": 0.2159347638582695,
"learning_rate": 1.638369302217687e-05,
"loss": 0.5919,
"step": 960
},
{
"epoch": 1.0604395604395604,
"grad_norm": 0.20846702806767928,
"learning_rate": 1.633435291081437e-05,
"loss": 0.5943,
"step": 965
},
{
"epoch": 1.065934065934066,
"grad_norm": 0.21732557354687126,
"learning_rate": 1.6284753900544384e-05,
"loss": 0.6074,
"step": 970
},
{
"epoch": 1.0714285714285714,
"grad_norm": 0.21491296524750328,
"learning_rate": 1.6234898018587336e-05,
"loss": 0.6033,
"step": 975
},
{
"epoch": 1.0769230769230769,
"grad_norm": 0.2420317609022985,
"learning_rate": 1.618478730266255e-05,
"loss": 0.6081,
"step": 980
},
{
"epoch": 1.0824175824175823,
"grad_norm": 0.21022308551794014,
"learning_rate": 1.6134423800904985e-05,
"loss": 0.5755,
"step": 985
},
{
"epoch": 1.0879120879120878,
"grad_norm": 0.20675117493646727,
"learning_rate": 1.6083809571781498e-05,
"loss": 0.5977,
"step": 990
},
{
"epoch": 1.0934065934065935,
"grad_norm": 0.21348585093868772,
"learning_rate": 1.6032946684006745e-05,
"loss": 0.5859,
"step": 995
},
{
"epoch": 1.098901098901099,
"grad_norm": 0.2840550425577152,
"learning_rate": 1.598183721645858e-05,
"loss": 0.5925,
"step": 1000
},
{
"epoch": 1.098901098901099,
"eval_loss": 0.6866068243980408,
"eval_runtime": 56.3212,
"eval_samples_per_second": 24.467,
"eval_steps_per_second": 0.107,
"step": 1000
},
{
"epoch": 1.1043956043956045,
"grad_norm": 0.21980107752404765,
"learning_rate": 1.5930483258093144e-05,
"loss": 0.5916,
"step": 1005
},
{
"epoch": 1.10989010989011,
"grad_norm": 0.22069775219844529,
"learning_rate": 1.5878886907859423e-05,
"loss": 0.6005,
"step": 1010
},
{
"epoch": 1.1153846153846154,
"grad_norm": 0.22181894571988076,
"learning_rate": 1.5827050274613512e-05,
"loss": 0.5862,
"step": 1015
},
{
"epoch": 1.120879120879121,
"grad_norm": 0.20906581970205632,
"learning_rate": 1.57749754770324e-05,
"loss": 0.5745,
"step": 1020
},
{
"epoch": 1.1263736263736264,
"grad_norm": 0.21467487180581052,
"learning_rate": 1.5722664643527362e-05,
"loss": 0.5796,
"step": 1025
},
{
"epoch": 1.1318681318681318,
"grad_norm": 0.2126168321569666,
"learning_rate": 1.567011991215699e-05,
"loss": 0.5842,
"step": 1030
},
{
"epoch": 1.1373626373626373,
"grad_norm": 0.21522455237738478,
"learning_rate": 1.561734343053979e-05,
"loss": 0.5979,
"step": 1035
},
{
"epoch": 1.1428571428571428,
"grad_norm": 0.22633801737454678,
"learning_rate": 1.5564337355766412e-05,
"loss": 0.5759,
"step": 1040
},
{
"epoch": 1.1483516483516483,
"grad_norm": 0.21358958977145795,
"learning_rate": 1.551110385431148e-05,
"loss": 0.5939,
"step": 1045
},
{
"epoch": 1.1538461538461537,
"grad_norm": 0.23012592316508443,
"learning_rate": 1.5457645101945046e-05,
"loss": 0.5844,
"step": 1050
},
{
"epoch": 1.1593406593406592,
"grad_norm": 0.22381921198752516,
"learning_rate": 1.540396328364367e-05,
"loss": 0.5952,
"step": 1055
},
{
"epoch": 1.164835164835165,
"grad_norm": 0.21581592323525847,
"learning_rate": 1.5350060593501086e-05,
"loss": 0.5964,
"step": 1060
},
{
"epoch": 1.1703296703296704,
"grad_norm": 0.21213106121775074,
"learning_rate": 1.5295939234638566e-05,
"loss": 0.5948,
"step": 1065
},
{
"epoch": 1.1758241758241759,
"grad_norm": 0.2148049280696334,
"learning_rate": 1.5241601419114842e-05,
"loss": 0.5764,
"step": 1070
},
{
"epoch": 1.1813186813186813,
"grad_norm": 0.22960383958999012,
"learning_rate": 1.5187049367835709e-05,
"loss": 0.6003,
"step": 1075
},
{
"epoch": 1.1868131868131868,
"grad_norm": 0.21916919776904648,
"learning_rate": 1.5132285310463243e-05,
"loss": 0.5821,
"step": 1080
},
{
"epoch": 1.1923076923076923,
"grad_norm": 0.20455516360752496,
"learning_rate": 1.507731148532468e-05,
"loss": 0.5725,
"step": 1085
},
{
"epoch": 1.1978021978021978,
"grad_norm": 0.21171630637088715,
"learning_rate": 1.5022130139320916e-05,
"loss": 0.6011,
"step": 1090
},
{
"epoch": 1.2032967032967032,
"grad_norm": 0.21663323232960344,
"learning_rate": 1.4966743527834691e-05,
"loss": 0.5824,
"step": 1095
},
{
"epoch": 1.2087912087912087,
"grad_norm": 0.21254281294030117,
"learning_rate": 1.4911153914638388e-05,
"loss": 0.585,
"step": 1100
},
{
"epoch": 1.2087912087912087,
"eval_loss": 0.6832450032234192,
"eval_runtime": 57.3698,
"eval_samples_per_second": 24.02,
"eval_steps_per_second": 0.105,
"step": 1100
},
{
"epoch": 1.2142857142857142,
"grad_norm": 0.21064353654883214,
"learning_rate": 1.4855363571801523e-05,
"loss": 0.5825,
"step": 1105
},
{
"epoch": 1.2197802197802199,
"grad_norm": 0.21183460637710103,
"learning_rate": 1.4799374779597866e-05,
"loss": 0.5845,
"step": 1110
},
{
"epoch": 1.2252747252747254,
"grad_norm": 0.21260336129469928,
"learning_rate": 1.474318982641225e-05,
"loss": 0.5677,
"step": 1115
},
{
"epoch": 1.2307692307692308,
"grad_norm": 0.2126511780407026,
"learning_rate": 1.4686811008647037e-05,
"loss": 0.5804,
"step": 1120
},
{
"epoch": 1.2362637362637363,
"grad_norm": 0.21779018291407423,
"learning_rate": 1.463024063062827e-05,
"loss": 0.5863,
"step": 1125
},
{
"epoch": 1.2417582417582418,
"grad_norm": 0.20654603994279122,
"learning_rate": 1.457348100451146e-05,
"loss": 0.594,
"step": 1130
},
{
"epoch": 1.2472527472527473,
"grad_norm": 0.2085523679075223,
"learning_rate": 1.4516534450187126e-05,
"loss": 0.5755,
"step": 1135
},
{
"epoch": 1.2527472527472527,
"grad_norm": 0.20261083379143977,
"learning_rate": 1.4459403295185933e-05,
"loss": 0.5898,
"step": 1140
},
{
"epoch": 1.2582417582417582,
"grad_norm": 0.21173700024568526,
"learning_rate": 1.4402089874583594e-05,
"loss": 0.5757,
"step": 1145
},
{
"epoch": 1.2637362637362637,
"grad_norm": 0.20789003864449884,
"learning_rate": 1.4344596530905412e-05,
"loss": 0.5956,
"step": 1150
},
{
"epoch": 1.2692307692307692,
"grad_norm": 0.21703411260373892,
"learning_rate": 1.4286925614030542e-05,
"loss": 0.5783,
"step": 1155
},
{
"epoch": 1.2747252747252746,
"grad_norm": 0.20821140503325464,
"learning_rate": 1.4229079481095949e-05,
"loss": 0.5698,
"step": 1160
},
{
"epoch": 1.2802197802197801,
"grad_norm": 0.20754015036739895,
"learning_rate": 1.4171060496400055e-05,
"loss": 0.5683,
"step": 1165
},
{
"epoch": 1.2857142857142856,
"grad_norm": 0.21632631037889694,
"learning_rate": 1.4112871031306118e-05,
"loss": 0.5722,
"step": 1170
},
{
"epoch": 1.2912087912087913,
"grad_norm": 0.21139448126540564,
"learning_rate": 1.4054513464145303e-05,
"loss": 0.5941,
"step": 1175
},
{
"epoch": 1.2967032967032968,
"grad_norm": 0.2005961523532221,
"learning_rate": 1.3995990180119478e-05,
"loss": 0.5783,
"step": 1180
},
{
"epoch": 1.3021978021978022,
"grad_norm": 0.22506737587889664,
"learning_rate": 1.3937303571203718e-05,
"loss": 0.5815,
"step": 1185
},
{
"epoch": 1.3076923076923077,
"grad_norm": 0.20776135745507532,
"learning_rate": 1.387845603604855e-05,
"loss": 0.5754,
"step": 1190
},
{
"epoch": 1.3131868131868132,
"grad_norm": 0.20550133841197799,
"learning_rate": 1.3819449979881907e-05,
"loss": 0.5784,
"step": 1195
},
{
"epoch": 1.3186813186813187,
"grad_norm": 0.20805118712132745,
"learning_rate": 1.3760287814410822e-05,
"loss": 0.5857,
"step": 1200
},
{
"epoch": 1.3186813186813187,
"eval_loss": 0.6798372864723206,
"eval_runtime": 56.9599,
"eval_samples_per_second": 24.192,
"eval_steps_per_second": 0.105,
"step": 1200
},
{
"epoch": 1.3241758241758241,
"grad_norm": 0.2139723251065959,
"learning_rate": 1.3700971957722861e-05,
"loss": 0.5755,
"step": 1205
},
{
"epoch": 1.3296703296703296,
"grad_norm": 0.20154696968263214,
"learning_rate": 1.3641504834187288e-05,
"loss": 0.5905,
"step": 1210
},
{
"epoch": 1.335164835164835,
"grad_norm": 0.21481447035842982,
"learning_rate": 1.3581888874355969e-05,
"loss": 0.5968,
"step": 1215
},
{
"epoch": 1.3406593406593408,
"grad_norm": 0.23044920349242504,
"learning_rate": 1.3522126514864047e-05,
"loss": 0.5788,
"step": 1220
},
{
"epoch": 1.3461538461538463,
"grad_norm": 0.19789912123306916,
"learning_rate": 1.346222019833033e-05,
"loss": 0.5808,
"step": 1225
},
{
"epoch": 1.3516483516483517,
"grad_norm": 0.21068837876070498,
"learning_rate": 1.3402172373257466e-05,
"loss": 0.5777,
"step": 1230
},
{
"epoch": 1.3571428571428572,
"grad_norm": 0.20881025023596286,
"learning_rate": 1.3341985493931877e-05,
"loss": 0.5918,
"step": 1235
},
{
"epoch": 1.3626373626373627,
"grad_norm": 0.21040201908308628,
"learning_rate": 1.3281662020323434e-05,
"loss": 0.5678,
"step": 1240
},
{
"epoch": 1.3681318681318682,
"grad_norm": 0.21648376396617017,
"learning_rate": 1.3221204417984907e-05,
"loss": 0.5816,
"step": 1245
},
{
"epoch": 1.3736263736263736,
"grad_norm": 0.23069884686246664,
"learning_rate": 1.3160615157951218e-05,
"loss": 0.5825,
"step": 1250
},
{
"epoch": 1.379120879120879,
"grad_norm": 0.21219224260123462,
"learning_rate": 1.3099896716638414e-05,
"loss": 0.5862,
"step": 1255
},
{
"epoch": 1.3846153846153846,
"grad_norm": 0.20937858254743488,
"learning_rate": 1.303905157574247e-05,
"loss": 0.5754,
"step": 1260
},
{
"epoch": 1.39010989010989,
"grad_norm": 0.287029581105537,
"learning_rate": 1.297808222213785e-05,
"loss": 0.5723,
"step": 1265
},
{
"epoch": 1.3956043956043955,
"grad_norm": 0.2115500774984686,
"learning_rate": 1.2916991147775867e-05,
"loss": 0.5903,
"step": 1270
},
{
"epoch": 1.401098901098901,
"grad_norm": 0.2360214862879323,
"learning_rate": 1.2855780849582828e-05,
"loss": 0.5995,
"step": 1275
},
{
"epoch": 1.4065934065934065,
"grad_norm": 0.20079252362217995,
"learning_rate": 1.2794453829357974e-05,
"loss": 0.5845,
"step": 1280
},
{
"epoch": 1.412087912087912,
"grad_norm": 0.21328822844172537,
"learning_rate": 1.2733012593671235e-05,
"loss": 0.5764,
"step": 1285
},
{
"epoch": 1.4175824175824177,
"grad_norm": 0.20595464274671632,
"learning_rate": 1.2671459653760781e-05,
"loss": 0.5722,
"step": 1290
},
{
"epoch": 1.4230769230769231,
"grad_norm": 0.21114157296745673,
"learning_rate": 1.2609797525430374e-05,
"loss": 0.5851,
"step": 1295
},
{
"epoch": 1.4285714285714286,
"grad_norm": 0.28130303512847965,
"learning_rate": 1.2548028728946548e-05,
"loss": 0.5736,
"step": 1300
},
{
"epoch": 1.4285714285714286,
"eval_loss": 0.6746197938919067,
"eval_runtime": 56.6881,
"eval_samples_per_second": 24.308,
"eval_steps_per_second": 0.106,
"step": 1300
},
{
"epoch": 1.434065934065934,
"grad_norm": 0.2112732742944607,
"learning_rate": 1.2486155788935599e-05,
"loss": 0.5856,
"step": 1305
},
{
"epoch": 1.4395604395604396,
"grad_norm": 0.21296297985278367,
"learning_rate": 1.24241812342804e-05,
"loss": 0.5741,
"step": 1310
},
{
"epoch": 1.445054945054945,
"grad_norm": 0.20477018149128742,
"learning_rate": 1.2362107598017037e-05,
"loss": 0.5747,
"step": 1315
},
{
"epoch": 1.4505494505494505,
"grad_norm": 0.20359961146660085,
"learning_rate": 1.2299937417231269e-05,
"loss": 0.5909,
"step": 1320
},
{
"epoch": 1.456043956043956,
"grad_norm": 0.19983003339561645,
"learning_rate": 1.2237673232954854e-05,
"loss": 0.5819,
"step": 1325
},
{
"epoch": 1.4615384615384617,
"grad_norm": 0.21585126952570063,
"learning_rate": 1.2175317590061676e-05,
"loss": 0.5744,
"step": 1330
},
{
"epoch": 1.4670329670329672,
"grad_norm": 0.2164749075408294,
"learning_rate": 1.2112873037163728e-05,
"loss": 0.5921,
"step": 1335
},
{
"epoch": 1.4725274725274726,
"grad_norm": 0.21214564202757996,
"learning_rate": 1.2050342126506958e-05,
"loss": 0.6007,
"step": 1340
},
{
"epoch": 1.478021978021978,
"grad_norm": 0.20961989493961736,
"learning_rate": 1.1987727413866936e-05,
"loss": 0.5922,
"step": 1345
},
{
"epoch": 1.4835164835164836,
"grad_norm": 0.2070781399263859,
"learning_rate": 1.1925031458444416e-05,
"loss": 0.5579,
"step": 1350
},
{
"epoch": 1.489010989010989,
"grad_norm": 0.21230956255901673,
"learning_rate": 1.1862256822760704e-05,
"loss": 0.583,
"step": 1355
},
{
"epoch": 1.4945054945054945,
"grad_norm": 0.20764551222516942,
"learning_rate": 1.1799406072552963e-05,
"loss": 0.5881,
"step": 1360
},
{
"epoch": 1.5,
"grad_norm": 0.2216686537947161,
"learning_rate": 1.1736481776669307e-05,
"loss": 0.5884,
"step": 1365
},
{
"epoch": 1.5054945054945055,
"grad_norm": 0.20026575192100984,
"learning_rate": 1.1673486506963824e-05,
"loss": 0.5789,
"step": 1370
},
{
"epoch": 1.510989010989011,
"grad_norm": 0.20025791445692404,
"learning_rate": 1.1610422838191473e-05,
"loss": 0.5757,
"step": 1375
},
{
"epoch": 1.5164835164835164,
"grad_norm": 0.21762293248064593,
"learning_rate": 1.1547293347902813e-05,
"loss": 0.595,
"step": 1380
},
{
"epoch": 1.521978021978022,
"grad_norm": 0.20522965991028572,
"learning_rate": 1.148410061633869e-05,
"loss": 0.5825,
"step": 1385
},
{
"epoch": 1.5274725274725274,
"grad_norm": 0.2007877147929741,
"learning_rate": 1.1420847226324746e-05,
"loss": 0.5699,
"step": 1390
},
{
"epoch": 1.5329670329670328,
"grad_norm": 0.2117099052435157,
"learning_rate": 1.135753576316588e-05,
"loss": 0.5814,
"step": 1395
},
{
"epoch": 1.5384615384615383,
"grad_norm": 0.26509052806023586,
"learning_rate": 1.1294168814540554e-05,
"loss": 0.5906,
"step": 1400
},
{
"epoch": 1.5384615384615383,
"eval_loss": 0.6723337769508362,
"eval_runtime": 55.2482,
"eval_samples_per_second": 24.942,
"eval_steps_per_second": 0.109,
"step": 1400
},
{
"epoch": 1.5439560439560438,
"grad_norm": 0.21156955434544839,
"learning_rate": 1.1230748970395056e-05,
"loss": 0.577,
"step": 1405
},
{
"epoch": 1.5494505494505495,
"grad_norm": 0.20751359958793425,
"learning_rate": 1.1167278822837621e-05,
"loss": 0.5655,
"step": 1410
},
{
"epoch": 1.554945054945055,
"grad_norm": 0.21010912643848495,
"learning_rate": 1.1103760966032497e-05,
"loss": 0.5657,
"step": 1415
},
{
"epoch": 1.5604395604395604,
"grad_norm": 0.20804119817991984,
"learning_rate": 1.1040197996093915e-05,
"loss": 0.5829,
"step": 1420
},
{
"epoch": 1.565934065934066,
"grad_norm": 0.200793527641334,
"learning_rate": 1.0976592510979982e-05,
"loss": 0.5695,
"step": 1425
},
{
"epoch": 1.5714285714285714,
"grad_norm": 0.21571689584942777,
"learning_rate": 1.0912947110386484e-05,
"loss": 0.5687,
"step": 1430
},
{
"epoch": 1.5769230769230769,
"grad_norm": 0.20881262462287056,
"learning_rate": 1.084926439564065e-05,
"loss": 0.5813,
"step": 1435
},
{
"epoch": 1.5824175824175826,
"grad_norm": 0.19724870907211803,
"learning_rate": 1.0785546969594813e-05,
"loss": 0.5695,
"step": 1440
},
{
"epoch": 1.587912087912088,
"grad_norm": 0.19796826932631745,
"learning_rate": 1.0721797436520044e-05,
"loss": 0.5745,
"step": 1445
},
{
"epoch": 1.5934065934065935,
"grad_norm": 0.20623299794109134,
"learning_rate": 1.0658018401999681e-05,
"loss": 0.5916,
"step": 1450
},
{
"epoch": 1.598901098901099,
"grad_norm": 0.22551529540554374,
"learning_rate": 1.0594212472822865e-05,
"loss": 0.5844,
"step": 1455
},
{
"epoch": 1.6043956043956045,
"grad_norm": 0.20013753224483868,
"learning_rate": 1.053038225687798e-05,
"loss": 0.5767,
"step": 1460
},
{
"epoch": 1.60989010989011,
"grad_norm": 0.20025644857315295,
"learning_rate": 1.0466530363046057e-05,
"loss": 0.57,
"step": 1465
},
{
"epoch": 1.6153846153846154,
"grad_norm": 0.22069359110995249,
"learning_rate": 1.0402659401094154e-05,
"loss": 0.5803,
"step": 1470
},
{
"epoch": 1.620879120879121,
"grad_norm": 0.2016457322498759,
"learning_rate": 1.033877198156868e-05,
"loss": 0.5711,
"step": 1475
},
{
"epoch": 1.6263736263736264,
"grad_norm": 0.20948058749001328,
"learning_rate": 1.0274870715688713e-05,
"loss": 0.5614,
"step": 1480
},
{
"epoch": 1.6318681318681318,
"grad_norm": 0.20613326525368722,
"learning_rate": 1.0210958215239249e-05,
"loss": 0.5617,
"step": 1485
},
{
"epoch": 1.6373626373626373,
"grad_norm": 0.20057743312747597,
"learning_rate": 1.0147037092464469e-05,
"loss": 0.5749,
"step": 1490
},
{
"epoch": 1.6428571428571428,
"grad_norm": 0.21757021973274904,
"learning_rate": 1.0083109959960974e-05,
"loss": 0.5719,
"step": 1495
},
{
"epoch": 1.6483516483516483,
"grad_norm": 0.1988420530692502,
"learning_rate": 1.0019179430570984e-05,
"loss": 0.569,
"step": 1500
},
{
"epoch": 1.6483516483516483,
"eval_loss": 0.6686215996742249,
"eval_runtime": 57.8791,
"eval_samples_per_second": 23.808,
"eval_steps_per_second": 0.104,
"step": 1500
},
{
"epoch": 1.6538461538461537,
"grad_norm": 0.21660113082980018,
"learning_rate": 9.955248117275566e-06,
"loss": 0.5584,
"step": 1505
},
{
"epoch": 1.6593406593406592,
"grad_norm": 0.23375649427094025,
"learning_rate": 9.891318633087831e-06,
"loss": 0.5515,
"step": 1510
},
{
"epoch": 1.6648351648351647,
"grad_norm": 0.20598295650184129,
"learning_rate": 9.827393590946116e-06,
"loss": 0.5591,
"step": 1515
},
{
"epoch": 1.6703296703296702,
"grad_norm": 0.19812798829418632,
"learning_rate": 9.763475603607215e-06,
"loss": 0.5589,
"step": 1520
},
{
"epoch": 1.6758241758241759,
"grad_norm": 0.20951557993767064,
"learning_rate": 9.699567283539567e-06,
"loss": 0.5777,
"step": 1525
},
{
"epoch": 1.6813186813186813,
"grad_norm": 0.20861568891956944,
"learning_rate": 9.635671242816503e-06,
"loss": 0.5793,
"step": 1530
},
{
"epoch": 1.6868131868131868,
"grad_norm": 0.2045724615591832,
"learning_rate": 9.571790093009445e-06,
"loss": 0.5507,
"step": 1535
},
{
"epoch": 1.6923076923076923,
"grad_norm": 0.2062400013339212,
"learning_rate": 9.50792644508122e-06,
"loss": 0.5859,
"step": 1540
},
{
"epoch": 1.6978021978021978,
"grad_norm": 0.20539716997080384,
"learning_rate": 9.44408290927929e-06,
"loss": 0.5883,
"step": 1545
},
{
"epoch": 1.7032967032967035,
"grad_norm": 0.2007069765651346,
"learning_rate": 9.380262095029113e-06,
"loss": 0.5433,
"step": 1550
},
{
"epoch": 1.708791208791209,
"grad_norm": 0.19988561870754676,
"learning_rate": 9.316466610827446e-06,
"loss": 0.5681,
"step": 1555
},
{
"epoch": 1.7142857142857144,
"grad_norm": 0.20674989840235833,
"learning_rate": 9.252699064135759e-06,
"loss": 0.56,
"step": 1560
},
{
"epoch": 1.7197802197802199,
"grad_norm": 0.19685538726018847,
"learning_rate": 9.188962061273664e-06,
"loss": 0.5762,
"step": 1565
},
{
"epoch": 1.7252747252747254,
"grad_norm": 0.20541519207163753,
"learning_rate": 9.125258207312365e-06,
"loss": 0.5637,
"step": 1570
},
{
"epoch": 1.7307692307692308,
"grad_norm": 0.1975839480015212,
"learning_rate": 9.061590105968208e-06,
"loss": 0.5529,
"step": 1575
},
{
"epoch": 1.7362637362637363,
"grad_norm": 0.22500565969016117,
"learning_rate": 8.997960359496248e-06,
"loss": 0.5736,
"step": 1580
},
{
"epoch": 1.7417582417582418,
"grad_norm": 0.19091706107610168,
"learning_rate": 8.934371568583893e-06,
"loss": 0.5683,
"step": 1585
},
{
"epoch": 1.7472527472527473,
"grad_norm": 0.2105931306233096,
"learning_rate": 8.8708263322446e-06,
"loss": 0.5751,
"step": 1590
},
{
"epoch": 1.7527472527472527,
"grad_norm": 0.20458297854035193,
"learning_rate": 8.807327247711667e-06,
"loss": 0.5873,
"step": 1595
},
{
"epoch": 1.7582417582417582,
"grad_norm": 0.19955511509412213,
"learning_rate": 8.743876910332057e-06,
"loss": 0.5756,
"step": 1600
},
{
"epoch": 1.7582417582417582,
"eval_loss": 0.6655026078224182,
"eval_runtime": 58.1146,
"eval_samples_per_second": 23.712,
"eval_steps_per_second": 0.103,
"step": 1600
},
{
"epoch": 1.7637362637362637,
"grad_norm": 0.20167066864184663,
"learning_rate": 8.680477913460339e-06,
"loss": 0.5625,
"step": 1605
},
{
"epoch": 1.7692307692307692,
"grad_norm": 0.20048559966418272,
"learning_rate": 8.617132848352672e-06,
"loss": 0.589,
"step": 1610
},
{
"epoch": 1.7747252747252746,
"grad_norm": 0.20227923458206726,
"learning_rate": 8.553844304060908e-06,
"loss": 0.5601,
"step": 1615
},
{
"epoch": 1.7802197802197801,
"grad_norm": 0.19698401864090306,
"learning_rate": 8.490614867326775e-06,
"loss": 0.5653,
"step": 1620
},
{
"epoch": 1.7857142857142856,
"grad_norm": 0.20175985199043966,
"learning_rate": 8.427447122476148e-06,
"loss": 0.576,
"step": 1625
},
{
"epoch": 1.791208791208791,
"grad_norm": 0.20649958059932677,
"learning_rate": 8.364343651313406e-06,
"loss": 0.5789,
"step": 1630
},
{
"epoch": 1.7967032967032965,
"grad_norm": 0.20362236096913583,
"learning_rate": 8.301307033015928e-06,
"loss": 0.5743,
"step": 1635
},
{
"epoch": 1.8021978021978022,
"grad_norm": 0.1986543134675767,
"learning_rate": 8.23833984402868e-06,
"loss": 0.5643,
"step": 1640
},
{
"epoch": 1.8076923076923077,
"grad_norm": 0.198913330885529,
"learning_rate": 8.175444657958875e-06,
"loss": 0.5735,
"step": 1645
},
{
"epoch": 1.8131868131868132,
"grad_norm": 0.19864368535876992,
"learning_rate": 8.112624045470834e-06,
"loss": 0.5717,
"step": 1650
},
{
"epoch": 1.8186813186813187,
"grad_norm": 0.19883717015134883,
"learning_rate": 8.04988057418088e-06,
"loss": 0.5557,
"step": 1655
},
{
"epoch": 1.8241758241758241,
"grad_norm": 0.20332693450451117,
"learning_rate": 7.987216808552409e-06,
"loss": 0.5736,
"step": 1660
},
{
"epoch": 1.8296703296703298,
"grad_norm": 0.2022305746327796,
"learning_rate": 7.924635309791065e-06,
"loss": 0.5605,
"step": 1665
},
{
"epoch": 1.8351648351648353,
"grad_norm": 0.20043133551344314,
"learning_rate": 7.862138635740078e-06,
"loss": 0.5509,
"step": 1670
},
{
"epoch": 1.8406593406593408,
"grad_norm": 0.22455898506641436,
"learning_rate": 7.799729340775688e-06,
"loss": 0.5869,
"step": 1675
},
{
"epoch": 1.8461538461538463,
"grad_norm": 0.19030565857505835,
"learning_rate": 7.73740997570278e-06,
"loss": 0.5657,
"step": 1680
},
{
"epoch": 1.8516483516483517,
"grad_norm": 0.20815080209673342,
"learning_rate": 7.675183087650592e-06,
"loss": 0.566,
"step": 1685
},
{
"epoch": 1.8571428571428572,
"grad_norm": 0.20412451434740683,
"learning_rate": 7.613051219968624e-06,
"loss": 0.5571,
"step": 1690
},
{
"epoch": 1.8626373626373627,
"grad_norm": 0.20467604912515247,
"learning_rate": 7.551016912122692e-06,
"loss": 0.5677,
"step": 1695
},
{
"epoch": 1.8681318681318682,
"grad_norm": 0.2049947541178511,
"learning_rate": 7.489082699591128e-06,
"loss": 0.545,
"step": 1700
},
{
"epoch": 1.8681318681318682,
"eval_loss": 0.6621970534324646,
"eval_runtime": 53.6631,
"eval_samples_per_second": 25.679,
"eval_steps_per_second": 0.112,
"step": 1700
},
{
"epoch": 1.8736263736263736,
"grad_norm": 0.20749354283076393,
"learning_rate": 7.4272511137611405e-06,
"loss": 0.5875,
"step": 1705
},
{
"epoch": 1.879120879120879,
"grad_norm": 0.19337165187460637,
"learning_rate": 7.3655246818253626e-06,
"loss": 0.5633,
"step": 1710
},
{
"epoch": 1.8846153846153846,
"grad_norm": 0.20317650293507142,
"learning_rate": 7.303905926678565e-06,
"loss": 0.5625,
"step": 1715
},
{
"epoch": 1.89010989010989,
"grad_norm": 0.20166481148644033,
"learning_rate": 7.242397366814516e-06,
"loss": 0.5444,
"step": 1720
},
{
"epoch": 1.8956043956043955,
"grad_norm": 0.20130249873133135,
"learning_rate": 7.181001516223074e-06,
"loss": 0.5501,
"step": 1725
},
{
"epoch": 1.901098901098901,
"grad_norm": 0.20104391428043816,
"learning_rate": 7.1197208842874175e-06,
"loss": 0.5568,
"step": 1730
},
{
"epoch": 1.9065934065934065,
"grad_norm": 0.19669831797494047,
"learning_rate": 7.058557975681488e-06,
"loss": 0.5691,
"step": 1735
},
{
"epoch": 1.912087912087912,
"grad_norm": 0.19994455544845724,
"learning_rate": 6.997515290267611e-06,
"loss": 0.5641,
"step": 1740
},
{
"epoch": 1.9175824175824174,
"grad_norm": 0.2022020053476063,
"learning_rate": 6.936595322994328e-06,
"loss": 0.5625,
"step": 1745
},
{
"epoch": 1.9230769230769231,
"grad_norm": 0.20426025496672773,
"learning_rate": 6.8758005637944245e-06,
"loss": 0.5662,
"step": 1750
},
{
"epoch": 1.9285714285714286,
"grad_norm": 0.19660653769715136,
"learning_rate": 6.815133497483157e-06,
"loss": 0.575,
"step": 1755
},
{
"epoch": 1.934065934065934,
"grad_norm": 0.2091108562276251,
"learning_rate": 6.754596603656687e-06,
"loss": 0.5668,
"step": 1760
},
{
"epoch": 1.9395604395604396,
"grad_norm": 0.1990872798241454,
"learning_rate": 6.694192356590743e-06,
"loss": 0.5677,
"step": 1765
},
{
"epoch": 1.945054945054945,
"grad_norm": 0.2035184385391121,
"learning_rate": 6.633923225139498e-06,
"loss": 0.5848,
"step": 1770
},
{
"epoch": 1.9505494505494505,
"grad_norm": 0.2087152884886787,
"learning_rate": 6.573791672634638e-06,
"loss": 0.5779,
"step": 1775
},
{
"epoch": 1.9560439560439562,
"grad_norm": 0.19903589160386567,
"learning_rate": 6.513800156784709e-06,
"loss": 0.5545,
"step": 1780
},
{
"epoch": 1.9615384615384617,
"grad_norm": 0.1988257810600432,
"learning_rate": 6.453951129574644e-06,
"loss": 0.5591,
"step": 1785
},
{
"epoch": 1.9670329670329672,
"grad_norm": 0.2440643065747747,
"learning_rate": 6.394247037165559e-06,
"loss": 0.5575,
"step": 1790
},
{
"epoch": 1.9725274725274726,
"grad_norm": 0.19915770600503882,
"learning_rate": 6.3346903197947564e-06,
"loss": 0.5713,
"step": 1795
},
{
"epoch": 1.978021978021978,
"grad_norm": 0.19824070469872943,
"learning_rate": 6.275283411676008e-06,
"loss": 0.5505,
"step": 1800
},
{
"epoch": 1.978021978021978,
"eval_loss": 0.6606090068817139,
"eval_runtime": 57.2329,
"eval_samples_per_second": 24.077,
"eval_steps_per_second": 0.105,
"step": 1800
},
{
"epoch": 1.9835164835164836,
"grad_norm": 0.20291412687117444,
"learning_rate": 6.216028740900042e-06,
"loss": 0.5715,
"step": 1805
},
{
"epoch": 1.989010989010989,
"grad_norm": 0.1992635144672437,
"learning_rate": 6.1569287293353274e-06,
"loss": 0.5648,
"step": 1810
},
{
"epoch": 1.9945054945054945,
"grad_norm": 0.20133042378281404,
"learning_rate": 6.097985792529055e-06,
"loss": 0.5718,
"step": 1815
},
{
"epoch": 2.0,
"grad_norm": 0.18939749792824842,
"learning_rate": 6.039202339608432e-06,
"loss": 0.5509,
"step": 1820
},
{
"epoch": 2.0054945054945055,
"grad_norm": 0.21548294059114115,
"learning_rate": 5.980580773182214e-06,
"loss": 0.5084,
"step": 1825
},
{
"epoch": 2.010989010989011,
"grad_norm": 0.19938738432640987,
"learning_rate": 5.922123489242499e-06,
"loss": 0.5162,
"step": 1830
},
{
"epoch": 2.0164835164835164,
"grad_norm": 0.2027789624856937,
"learning_rate": 5.8638328770667905e-06,
"loss": 0.5119,
"step": 1835
},
{
"epoch": 2.021978021978022,
"grad_norm": 0.20197197450501087,
"learning_rate": 5.805711319120358e-06,
"loss": 0.5093,
"step": 1840
},
{
"epoch": 2.0274725274725274,
"grad_norm": 0.20124091857994694,
"learning_rate": 5.747761190958859e-06,
"loss": 0.5118,
"step": 1845
},
{
"epoch": 2.032967032967033,
"grad_norm": 0.2060214343274437,
"learning_rate": 5.689984861131221e-06,
"loss": 0.529,
"step": 1850
},
{
"epoch": 2.0384615384615383,
"grad_norm": 0.1949171402387555,
"learning_rate": 5.632384691082874e-06,
"loss": 0.5038,
"step": 1855
},
{
"epoch": 2.043956043956044,
"grad_norm": 0.2065371272676731,
"learning_rate": 5.5749630350592e-06,
"loss": 0.524,
"step": 1860
},
{
"epoch": 2.0494505494505493,
"grad_norm": 0.20211687410000778,
"learning_rate": 5.517722240009319e-06,
"loss": 0.5249,
"step": 1865
},
{
"epoch": 2.0549450549450547,
"grad_norm": 0.21155781039557575,
"learning_rate": 5.460664645490172e-06,
"loss": 0.5222,
"step": 1870
},
{
"epoch": 2.0604395604395602,
"grad_norm": 0.20557792586063273,
"learning_rate": 5.403792583570884e-06,
"loss": 0.5122,
"step": 1875
},
{
"epoch": 2.065934065934066,
"grad_norm": 0.20892091915093894,
"learning_rate": 5.347108378737469e-06,
"loss": 0.5209,
"step": 1880
},
{
"epoch": 2.0714285714285716,
"grad_norm": 0.21130418206749751,
"learning_rate": 5.290614347797802e-06,
"loss": 0.4999,
"step": 1885
},
{
"epoch": 2.076923076923077,
"grad_norm": 0.20265157501265516,
"learning_rate": 5.234312799786921e-06,
"loss": 0.5113,
"step": 1890
},
{
"epoch": 2.0824175824175826,
"grad_norm": 0.2380811857415137,
"learning_rate": 5.1782060358726885e-06,
"loss": 0.5137,
"step": 1895
},
{
"epoch": 2.087912087912088,
"grad_norm": 0.20021684752472454,
"learning_rate": 5.122296349261695e-06,
"loss": 0.5149,
"step": 1900
},
{
"epoch": 2.087912087912088,
"eval_loss": 0.6648159027099609,
"eval_runtime": 57.1476,
"eval_samples_per_second": 24.113,
"eval_steps_per_second": 0.105,
"step": 1900
},
{
"epoch": 2.0934065934065935,
"grad_norm": 0.20298609856841238,
"learning_rate": 5.066586025105558e-06,
"loss": 0.527,
"step": 1905
},
{
"epoch": 2.098901098901099,
"grad_norm": 0.2035387697050701,
"learning_rate": 5.011077340407509e-06,
"loss": 0.5055,
"step": 1910
},
{
"epoch": 2.1043956043956045,
"grad_norm": 0.19786708962281988,
"learning_rate": 4.955772563929334e-06,
"loss": 0.5151,
"step": 1915
},
{
"epoch": 2.10989010989011,
"grad_norm": 0.20387216209111395,
"learning_rate": 4.900673956098644e-06,
"loss": 0.5049,
"step": 1920
},
{
"epoch": 2.1153846153846154,
"grad_norm": 0.20554401944132458,
"learning_rate": 4.845783768916482e-06,
"loss": 0.519,
"step": 1925
},
{
"epoch": 2.120879120879121,
"grad_norm": 0.2111896896707393,
"learning_rate": 4.79110424586528e-06,
"loss": 0.5193,
"step": 1930
},
{
"epoch": 2.1263736263736264,
"grad_norm": 0.20409361712091278,
"learning_rate": 4.736637621817176e-06,
"loss": 0.5134,
"step": 1935
},
{
"epoch": 2.131868131868132,
"grad_norm": 0.20305198178835793,
"learning_rate": 4.682386122942649e-06,
"loss": 0.5338,
"step": 1940
},
{
"epoch": 2.1373626373626373,
"grad_norm": 0.20969960596506004,
"learning_rate": 4.628351966619531e-06,
"loss": 0.5285,
"step": 1945
},
{
"epoch": 2.142857142857143,
"grad_norm": 0.2016497737059224,
"learning_rate": 4.5745373613424075e-06,
"loss": 0.5196,
"step": 1950
},
{
"epoch": 2.1483516483516483,
"grad_norm": 0.20780959230200222,
"learning_rate": 4.520944506632314e-06,
"loss": 0.5102,
"step": 1955
},
{
"epoch": 2.1538461538461537,
"grad_norm": 0.20807222664357544,
"learning_rate": 4.467575592946865e-06,
"loss": 0.5325,
"step": 1960
},
{
"epoch": 2.159340659340659,
"grad_norm": 0.2167502541161845,
"learning_rate": 4.414432801590703e-06,
"loss": 0.5078,
"step": 1965
},
{
"epoch": 2.1648351648351647,
"grad_norm": 0.21490727997644427,
"learning_rate": 4.361518304626366e-06,
"loss": 0.5407,
"step": 1970
},
{
"epoch": 2.17032967032967,
"grad_norm": 0.21734918740860756,
"learning_rate": 4.308834264785483e-06,
"loss": 0.5141,
"step": 1975
},
{
"epoch": 2.1758241758241756,
"grad_norm": 0.1991788993982973,
"learning_rate": 4.256382835380421e-06,
"loss": 0.5069,
"step": 1980
},
{
"epoch": 2.181318681318681,
"grad_norm": 0.2143454079537833,
"learning_rate": 4.204166160216216e-06,
"loss": 0.5249,
"step": 1985
},
{
"epoch": 2.186813186813187,
"grad_norm": 0.19879558037154738,
"learning_rate": 4.1521863735030065e-06,
"loss": 0.4902,
"step": 1990
},
{
"epoch": 2.1923076923076925,
"grad_norm": 0.20002933623961325,
"learning_rate": 4.100445599768774e-06,
"loss": 0.5051,
"step": 1995
},
{
"epoch": 2.197802197802198,
"grad_norm": 0.20285409795219067,
"learning_rate": 4.048945953772504e-06,
"loss": 0.5234,
"step": 2000
},
{
"epoch": 2.197802197802198,
"eval_loss": 0.6638170480728149,
"eval_runtime": 57.5818,
"eval_samples_per_second": 23.931,
"eval_steps_per_second": 0.104,
"step": 2000
},
{
"epoch": 2.2032967032967035,
"grad_norm": 0.2018945000710767,
"learning_rate": 3.99768954041778e-06,
"loss": 0.5071,
"step": 2005
},
{
"epoch": 2.208791208791209,
"grad_norm": 0.20646923773133413,
"learning_rate": 3.946678454666719e-06,
"loss": 0.4988,
"step": 2010
},
{
"epoch": 2.2142857142857144,
"grad_norm": 0.20067691812172808,
"learning_rate": 3.89591478145437e-06,
"loss": 0.5182,
"step": 2015
},
{
"epoch": 2.21978021978022,
"grad_norm": 0.21205391255701184,
"learning_rate": 3.845400595603482e-06,
"loss": 0.5307,
"step": 2020
},
{
"epoch": 2.2252747252747254,
"grad_norm": 0.20460875351302696,
"learning_rate": 3.79513796173971e-06,
"loss": 0.5031,
"step": 2025
},
{
"epoch": 2.230769230769231,
"grad_norm": 0.1973070576811456,
"learning_rate": 3.745128934207225e-06,
"loss": 0.5284,
"step": 2030
},
{
"epoch": 2.2362637362637363,
"grad_norm": 0.21339279534660974,
"learning_rate": 3.695375556984764e-06,
"loss": 0.5252,
"step": 2035
},
{
"epoch": 2.241758241758242,
"grad_norm": 0.20247153184924946,
"learning_rate": 3.6458798636020477e-06,
"loss": 0.5128,
"step": 2040
},
{
"epoch": 2.2472527472527473,
"grad_norm": 0.19749668386741256,
"learning_rate": 3.59664387705672e-06,
"loss": 0.5062,
"step": 2045
},
{
"epoch": 2.2527472527472527,
"grad_norm": 0.21402795878492867,
"learning_rate": 3.5476696097316253e-06,
"loss": 0.5109,
"step": 2050
},
{
"epoch": 2.258241758241758,
"grad_norm": 0.2077870859543597,
"learning_rate": 3.4989590633125583e-06,
"loss": 0.5274,
"step": 2055
},
{
"epoch": 2.2637362637362637,
"grad_norm": 0.2660838503313638,
"learning_rate": 3.450514228706482e-06,
"loss": 0.5059,
"step": 2060
},
{
"epoch": 2.269230769230769,
"grad_norm": 0.20089637585406683,
"learning_rate": 3.4023370859601192e-06,
"loss": 0.5044,
"step": 2065
},
{
"epoch": 2.2747252747252746,
"grad_norm": 0.19730583864844892,
"learning_rate": 3.3544296041790457e-06,
"loss": 0.5002,
"step": 2070
},
{
"epoch": 2.28021978021978,
"grad_norm": 0.202178274091821,
"learning_rate": 3.3067937414471986e-06,
"loss": 0.5244,
"step": 2075
},
{
"epoch": 2.2857142857142856,
"grad_norm": 0.19886623288328378,
"learning_rate": 3.2594314447468457e-06,
"loss": 0.5115,
"step": 2080
},
{
"epoch": 2.291208791208791,
"grad_norm": 0.20284655112337058,
"learning_rate": 3.2123446498790214e-06,
"loss": 0.5108,
"step": 2085
},
{
"epoch": 2.2967032967032965,
"grad_norm": 0.21307262525015022,
"learning_rate": 3.1655352813843886e-06,
"loss": 0.5205,
"step": 2090
},
{
"epoch": 2.302197802197802,
"grad_norm": 0.2016117227949298,
"learning_rate": 3.1190052524645752e-06,
"loss": 0.5036,
"step": 2095
},
{
"epoch": 2.3076923076923075,
"grad_norm": 0.20155504502839525,
"learning_rate": 3.0727564649040066e-06,
"loss": 0.5239,
"step": 2100
},
{
"epoch": 2.3076923076923075,
"eval_loss": 0.6632149815559387,
"eval_runtime": 55.6194,
"eval_samples_per_second": 24.776,
"eval_steps_per_second": 0.108,
"step": 2100
},
{
"epoch": 2.313186813186813,
"grad_norm": 0.21425188656541483,
"learning_rate": 3.0267908089921438e-06,
"loss": 0.5151,
"step": 2105
},
{
"epoch": 2.3186813186813184,
"grad_norm": 0.20081101739821433,
"learning_rate": 2.9811101634462414e-06,
"loss": 0.5147,
"step": 2110
},
{
"epoch": 2.3241758241758244,
"grad_norm": 0.20232270966888258,
"learning_rate": 2.93571639533455e-06,
"loss": 0.522,
"step": 2115
},
{
"epoch": 2.32967032967033,
"grad_norm": 0.20125405245316808,
"learning_rate": 2.8906113600000153e-06,
"loss": 0.5209,
"step": 2120
},
{
"epoch": 2.3351648351648353,
"grad_norm": 0.20454677450030168,
"learning_rate": 2.8457969009844354e-06,
"loss": 0.5131,
"step": 2125
},
{
"epoch": 2.340659340659341,
"grad_norm": 0.2068065615757479,
"learning_rate": 2.8012748499531195e-06,
"loss": 0.5268,
"step": 2130
},
{
"epoch": 2.3461538461538463,
"grad_norm": 0.20463770808980064,
"learning_rate": 2.7570470266200177e-06,
"loss": 0.5269,
"step": 2135
},
{
"epoch": 2.3516483516483517,
"grad_norm": 0.20997582101579929,
"learning_rate": 2.713115238673356e-06,
"loss": 0.5247,
"step": 2140
},
{
"epoch": 2.357142857142857,
"grad_norm": 0.20156173202574512,
"learning_rate": 2.669481281701739e-06,
"loss": 0.5039,
"step": 2145
},
{
"epoch": 2.3626373626373627,
"grad_norm": 0.19654077766547687,
"learning_rate": 2.626146939120757e-06,
"loss": 0.5185,
"step": 2150
},
{
"epoch": 2.368131868131868,
"grad_norm": 0.20382023726525478,
"learning_rate": 2.5831139821001184e-06,
"loss": 0.5259,
"step": 2155
},
{
"epoch": 2.3736263736263736,
"grad_norm": 0.20680053202116233,
"learning_rate": 2.5403841694912333e-06,
"loss": 0.513,
"step": 2160
},
{
"epoch": 2.379120879120879,
"grad_norm": 0.1941492774491302,
"learning_rate": 2.497959247755335e-06,
"loss": 0.5239,
"step": 2165
},
{
"epoch": 2.3846153846153846,
"grad_norm": 0.20405640437952138,
"learning_rate": 2.455840950892099e-06,
"loss": 0.5156,
"step": 2170
},
{
"epoch": 2.39010989010989,
"grad_norm": 0.2843212523020798,
"learning_rate": 2.414031000368767e-06,
"loss": 0.5034,
"step": 2175
},
{
"epoch": 2.3956043956043955,
"grad_norm": 0.20197161260199611,
"learning_rate": 2.372531105049789e-06,
"loss": 0.5003,
"step": 2180
},
{
"epoch": 2.401098901098901,
"grad_norm": 0.20727591825649075,
"learning_rate": 2.331342961126988e-06,
"loss": 0.5079,
"step": 2185
},
{
"epoch": 2.4065934065934065,
"grad_norm": 0.2054230166031501,
"learning_rate": 2.290468252050204e-06,
"loss": 0.5034,
"step": 2190
},
{
"epoch": 2.412087912087912,
"grad_norm": 0.2011557586955495,
"learning_rate": 2.2499086484585255e-06,
"loss": 0.5023,
"step": 2195
},
{
"epoch": 2.4175824175824174,
"grad_norm": 0.2043529990444942,
"learning_rate": 2.2096658081119793e-06,
"loss": 0.5142,
"step": 2200
},
{
"epoch": 2.4175824175824174,
"eval_loss": 0.662318229675293,
"eval_runtime": 56.3794,
"eval_samples_per_second": 24.442,
"eval_steps_per_second": 0.106,
"step": 2200
},
{
"epoch": 2.423076923076923,
"grad_norm": 0.2007104518256576,
"learning_rate": 2.1697413758237785e-06,
"loss": 0.5068,
"step": 2205
},
{
"epoch": 2.4285714285714284,
"grad_norm": 0.20519816271013885,
"learning_rate": 2.130136983393112e-06,
"loss": 0.5153,
"step": 2210
},
{
"epoch": 2.4340659340659343,
"grad_norm": 0.19385746620653593,
"learning_rate": 2.0908542495384276e-06,
"loss": 0.4915,
"step": 2215
},
{
"epoch": 2.4395604395604398,
"grad_norm": 0.20504945100422145,
"learning_rate": 2.051894779831286e-06,
"loss": 0.5117,
"step": 2220
},
{
"epoch": 2.4450549450549453,
"grad_norm": 0.20350739624315656,
"learning_rate": 2.0132601666307295e-06,
"loss": 0.5118,
"step": 2225
},
{
"epoch": 2.4505494505494507,
"grad_norm": 0.2038803889957168,
"learning_rate": 1.9749519890182035e-06,
"loss": 0.5144,
"step": 2230
},
{
"epoch": 2.456043956043956,
"grad_norm": 0.20292180332730542,
"learning_rate": 1.936971812733012e-06,
"loss": 0.513,
"step": 2235
},
{
"epoch": 2.4615384615384617,
"grad_norm": 0.19993755435515206,
"learning_rate": 1.8993211901083353e-06,
"loss": 0.5145,
"step": 2240
},
{
"epoch": 2.467032967032967,
"grad_norm": 0.20670573681391916,
"learning_rate": 1.8620016600077516e-06,
"loss": 0.5159,
"step": 2245
},
{
"epoch": 2.4725274725274726,
"grad_norm": 0.19479382540898788,
"learning_rate": 1.8250147477623836e-06,
"loss": 0.5068,
"step": 2250
},
{
"epoch": 2.478021978021978,
"grad_norm": 0.20005895519952888,
"learning_rate": 1.7883619651085194e-06,
"loss": 0.5165,
"step": 2255
},
{
"epoch": 2.4835164835164836,
"grad_norm": 0.2009142357322209,
"learning_rate": 1.7520448101258325e-06,
"loss": 0.4934,
"step": 2260
},
{
"epoch": 2.489010989010989,
"grad_norm": 0.20615350297111454,
"learning_rate": 1.716064767176172e-06,
"loss": 0.4966,
"step": 2265
},
{
"epoch": 2.4945054945054945,
"grad_norm": 0.19574790617423804,
"learning_rate": 1.6804233068428678e-06,
"loss": 0.4991,
"step": 2270
},
{
"epoch": 2.5,
"grad_norm": 0.1968274274952375,
"learning_rate": 1.6451218858706374e-06,
"loss": 0.5132,
"step": 2275
},
{
"epoch": 2.5054945054945055,
"grad_norm": 0.19627672556133136,
"learning_rate": 1.6101619471060415e-06,
"loss": 0.4986,
"step": 2280
},
{
"epoch": 2.510989010989011,
"grad_norm": 0.232139426765921,
"learning_rate": 1.5755449194385164e-06,
"loss": 0.5119,
"step": 2285
},
{
"epoch": 2.5164835164835164,
"grad_norm": 0.20274183634685633,
"learning_rate": 1.5412722177419658e-06,
"loss": 0.5023,
"step": 2290
},
{
"epoch": 2.521978021978022,
"grad_norm": 0.1991951421133761,
"learning_rate": 1.5073452428169444e-06,
"loss": 0.4949,
"step": 2295
},
{
"epoch": 2.5274725274725274,
"grad_norm": 0.19379452216149648,
"learning_rate": 1.4737653813333774e-06,
"loss": 0.5086,
"step": 2300
},
{
"epoch": 2.5274725274725274,
"eval_loss": 0.6616050004959106,
"eval_runtime": 55.0037,
"eval_samples_per_second": 25.053,
"eval_steps_per_second": 0.109,
"step": 2300
},
{
"epoch": 2.532967032967033,
"grad_norm": 0.21494093339800838,
"learning_rate": 1.4405340057739203e-06,
"loss": 0.5233,
"step": 2305
},
{
"epoch": 2.5384615384615383,
"grad_norm": 0.20790725753107414,
"learning_rate": 1.407652474377832e-06,
"loss": 0.5262,
"step": 2310
},
{
"epoch": 2.543956043956044,
"grad_norm": 0.23490670194522875,
"learning_rate": 1.3751221310854778e-06,
"loss": 0.5064,
"step": 2315
},
{
"epoch": 2.5494505494505493,
"grad_norm": 0.2045871257636731,
"learning_rate": 1.3429443054833913e-06,
"loss": 0.5022,
"step": 2320
},
{
"epoch": 2.5549450549450547,
"grad_norm": 0.20260454811288686,
"learning_rate": 1.311120312749935e-06,
"loss": 0.5001,
"step": 2325
},
{
"epoch": 2.5604395604395602,
"grad_norm": 0.20417611729894805,
"learning_rate": 1.2796514536015492e-06,
"loss": 0.5051,
"step": 2330
},
{
"epoch": 2.5659340659340657,
"grad_norm": 0.20980337074664027,
"learning_rate": 1.2485390142395793e-06,
"loss": 0.5191,
"step": 2335
},
{
"epoch": 2.571428571428571,
"grad_norm": 0.20318999346422076,
"learning_rate": 1.2177842662977136e-06,
"loss": 0.5247,
"step": 2340
},
{
"epoch": 2.5769230769230766,
"grad_norm": 0.19629810633248806,
"learning_rate": 1.1873884667900125e-06,
"loss": 0.5117,
"step": 2345
},
{
"epoch": 2.5824175824175826,
"grad_norm": 0.20122953995800785,
"learning_rate": 1.1573528580595195e-06,
"loss": 0.5022,
"step": 2350
},
{
"epoch": 2.587912087912088,
"grad_norm": 0.19407168789538073,
"learning_rate": 1.1276786677274866e-06,
"loss": 0.5018,
"step": 2355
},
{
"epoch": 2.5934065934065935,
"grad_norm": 0.20866298600964597,
"learning_rate": 1.0983671086432146e-06,
"loss": 0.5131,
"step": 2360
},
{
"epoch": 2.598901098901099,
"grad_norm": 0.19859110363436613,
"learning_rate": 1.069419378834461e-06,
"loss": 0.5072,
"step": 2365
},
{
"epoch": 2.6043956043956045,
"grad_norm": 0.1972544497689168,
"learning_rate": 1.040836661458482e-06,
"loss": 0.5121,
"step": 2370
},
{
"epoch": 2.60989010989011,
"grad_norm": 0.2229095733014629,
"learning_rate": 1.0126201247536783e-06,
"loss": 0.5023,
"step": 2375
},
{
"epoch": 2.6153846153846154,
"grad_norm": 0.2078898580527157,
"learning_rate": 9.8477092199184e-07,
"loss": 0.4999,
"step": 2380
},
{
"epoch": 2.620879120879121,
"grad_norm": 0.20385043702217534,
"learning_rate": 9.57290191431013e-07,
"loss": 0.4973,
"step": 2385
},
{
"epoch": 2.6263736263736264,
"grad_norm": 0.20024077890781708,
"learning_rate": 9.301790562689794e-07,
"loss": 0.5115,
"step": 2390
},
{
"epoch": 2.631868131868132,
"grad_norm": 0.20229645144805108,
"learning_rate": 9.034386245973359e-07,
"loss": 0.518,
"step": 2395
},
{
"epoch": 2.6373626373626373,
"grad_norm": 0.20806472260190578,
"learning_rate": 8.770699893562273e-07,
"loss": 0.4998,
"step": 2400
},
{
"epoch": 2.6373626373626373,
"eval_loss": 0.6603860259056091,
"eval_runtime": 55.4261,
"eval_samples_per_second": 24.862,
"eval_steps_per_second": 0.108,
"step": 2400
},
{
"epoch": 2.642857142857143,
"grad_norm": 0.20212616235290484,
"learning_rate": 8.510742282896545e-07,
"loss": 0.5176,
"step": 2405
},
{
"epoch": 2.6483516483516483,
"grad_norm": 0.19926286905249757,
"learning_rate": 8.254524039014289e-07,
"loss": 0.504,
"step": 2410
},
{
"epoch": 2.6538461538461537,
"grad_norm": 0.20491396527113304,
"learning_rate": 8.002055634117578e-07,
"loss": 0.5127,
"step": 2415
},
{
"epoch": 2.659340659340659,
"grad_norm": 0.1984394044529964,
"learning_rate": 7.753347387144294e-07,
"loss": 0.4961,
"step": 2420
},
{
"epoch": 2.6648351648351647,
"grad_norm": 0.20457728804801084,
"learning_rate": 7.508409463346389e-07,
"loss": 0.5166,
"step": 2425
},
{
"epoch": 2.67032967032967,
"grad_norm": 0.2053681782347483,
"learning_rate": 7.26725187387446e-07,
"loss": 0.4985,
"step": 2430
},
{
"epoch": 2.675824175824176,
"grad_norm": 0.21034306160224336,
"learning_rate": 7.029884475368542e-07,
"loss": 0.5286,
"step": 2435
},
{
"epoch": 2.6813186813186816,
"grad_norm": 0.21208091059236112,
"learning_rate": 6.796316969555205e-07,
"loss": 0.5135,
"step": 2440
},
{
"epoch": 2.686813186813187,
"grad_norm": 0.20061617557170722,
"learning_rate": 6.566558902851161e-07,
"loss": 0.5248,
"step": 2445
},
{
"epoch": 2.6923076923076925,
"grad_norm": 0.2077389312882985,
"learning_rate": 6.340619665972847e-07,
"loss": 0.5265,
"step": 2450
},
{
"epoch": 2.697802197802198,
"grad_norm": 0.1907623497336921,
"learning_rate": 6.118508493552866e-07,
"loss": 0.5048,
"step": 2455
},
{
"epoch": 2.7032967032967035,
"grad_norm": 0.19886591589245772,
"learning_rate": 5.900234463762367e-07,
"loss": 0.506,
"step": 2460
},
{
"epoch": 2.708791208791209,
"grad_norm": 0.19691007260870758,
"learning_rate": 5.685806497940027e-07,
"loss": 0.511,
"step": 2465
},
{
"epoch": 2.7142857142857144,
"grad_norm": 0.19864992732710893,
"learning_rate": 5.475233360227516e-07,
"loss": 0.5093,
"step": 2470
},
{
"epoch": 2.71978021978022,
"grad_norm": 0.20129564143877238,
"learning_rate": 5.268523657211188e-07,
"loss": 0.4973,
"step": 2475
},
{
"epoch": 2.7252747252747254,
"grad_norm": 0.19814396411057444,
"learning_rate": 5.065685837570312e-07,
"loss": 0.4985,
"step": 2480
},
{
"epoch": 2.730769230769231,
"grad_norm": 0.19965550111609479,
"learning_rate": 4.866728191731829e-07,
"loss": 0.5111,
"step": 2485
},
{
"epoch": 2.7362637362637363,
"grad_norm": 0.19457379885149514,
"learning_rate": 4.671658851531424e-07,
"loss": 0.5057,
"step": 2490
},
{
"epoch": 2.741758241758242,
"grad_norm": 0.19940679212384896,
"learning_rate": 4.480485789881217e-07,
"loss": 0.5127,
"step": 2495
},
{
"epoch": 2.7472527472527473,
"grad_norm": 0.19251317932524864,
"learning_rate": 4.293216820443891e-07,
"loss": 0.5029,
"step": 2500
},
{
"epoch": 2.7472527472527473,
"eval_loss": 0.6601957678794861,
"eval_runtime": 54.7865,
"eval_samples_per_second": 25.152,
"eval_steps_per_second": 0.11,
"step": 2500
},
{
"epoch": 2.7527472527472527,
"grad_norm": 0.20141031688990624,
"learning_rate": 4.109859597313237e-07,
"loss": 0.5199,
"step": 2505
},
{
"epoch": 2.758241758241758,
"grad_norm": 0.2031083447799231,
"learning_rate": 3.9304216147014853e-07,
"loss": 0.5058,
"step": 2510
},
{
"epoch": 2.7637362637362637,
"grad_norm": 0.20105232675523538,
"learning_rate": 3.7549102066328226e-07,
"loss": 0.512,
"step": 2515
},
{
"epoch": 2.769230769230769,
"grad_norm": 0.39484901944232603,
"learning_rate": 3.5833325466437697e-07,
"loss": 0.5012,
"step": 2520
},
{
"epoch": 2.7747252747252746,
"grad_norm": 0.19709914420778502,
"learning_rate": 3.4156956474898805e-07,
"loss": 0.5045,
"step": 2525
},
{
"epoch": 2.78021978021978,
"grad_norm": 0.20621710050936062,
"learning_rate": 3.2520063608592165e-07,
"loss": 0.5121,
"step": 2530
},
{
"epoch": 2.7857142857142856,
"grad_norm": 0.1991105530357979,
"learning_rate": 3.0922713770922155e-07,
"loss": 0.5203,
"step": 2535
},
{
"epoch": 2.791208791208791,
"grad_norm": 0.20012961338063315,
"learning_rate": 2.9364972249082747e-07,
"loss": 0.5184,
"step": 2540
},
{
"epoch": 2.7967032967032965,
"grad_norm": 0.19896717513784648,
"learning_rate": 2.7846902711389236e-07,
"loss": 0.5108,
"step": 2545
},
{
"epoch": 2.802197802197802,
"grad_norm": 0.1977548907547512,
"learning_rate": 2.636856720467573e-07,
"loss": 0.5208,
"step": 2550
},
{
"epoch": 2.8076923076923075,
"grad_norm": 0.20270402323570907,
"learning_rate": 2.493002615175977e-07,
"loss": 0.5006,
"step": 2555
},
{
"epoch": 2.813186813186813,
"grad_norm": 0.1942837253435945,
"learning_rate": 2.3531338348971366e-07,
"loss": 0.5119,
"step": 2560
},
{
"epoch": 2.8186813186813184,
"grad_norm": 0.20474339797776248,
"learning_rate": 2.217256096375131e-07,
"loss": 0.5176,
"step": 2565
},
{
"epoch": 2.824175824175824,
"grad_norm": 0.2010552477495884,
"learning_rate": 2.0853749532314006e-07,
"loss": 0.5073,
"step": 2570
},
{
"epoch": 2.82967032967033,
"grad_norm": 0.2035067000005909,
"learning_rate": 1.9574957957377294e-07,
"loss": 0.5096,
"step": 2575
},
{
"epoch": 2.8351648351648353,
"grad_norm": 0.20125633333431672,
"learning_rate": 1.8336238505959892e-07,
"loss": 0.5121,
"step": 2580
},
{
"epoch": 2.840659340659341,
"grad_norm": 0.20332908603793057,
"learning_rate": 1.7137641807244754e-07,
"loss": 0.5125,
"step": 2585
},
{
"epoch": 2.8461538461538463,
"grad_norm": 0.19665455030518153,
"learning_rate": 1.5979216850509848e-07,
"loss": 0.4987,
"step": 2590
},
{
"epoch": 2.8516483516483517,
"grad_norm": 0.1966270172986848,
"learning_rate": 1.4861010983126202e-07,
"loss": 0.4874,
"step": 2595
},
{
"epoch": 2.857142857142857,
"grad_norm": 0.19834917582035766,
"learning_rate": 1.3783069908621772e-07,
"loss": 0.5146,
"step": 2600
},
{
"epoch": 2.857142857142857,
"eval_loss": 0.6599460244178772,
"eval_runtime": 56.5256,
"eval_samples_per_second": 24.378,
"eval_steps_per_second": 0.106,
"step": 2600
},
{
"epoch": 2.8626373626373627,
"grad_norm": 0.19885509465905268,
"learning_rate": 1.274543768481451e-07,
"loss": 0.5192,
"step": 2605
},
{
"epoch": 2.868131868131868,
"grad_norm": 0.19541257662283346,
"learning_rate": 1.1748156722011128e-07,
"loss": 0.4926,
"step": 2610
},
{
"epoch": 2.8736263736263736,
"grad_norm": 0.19388744876156347,
"learning_rate": 1.0791267781273263e-07,
"loss": 0.5094,
"step": 2615
},
{
"epoch": 2.879120879120879,
"grad_norm": 0.1976970112287548,
"learning_rate": 9.874809972752697e-08,
"loss": 0.5156,
"step": 2620
},
{
"epoch": 2.8846153846153846,
"grad_norm": 0.20191971179188523,
"learning_rate": 8.99882075409153e-08,
"loss": 0.5176,
"step": 2625
},
{
"epoch": 2.89010989010989,
"grad_norm": 0.20009589256600918,
"learning_rate": 8.16333592889207e-08,
"loss": 0.5135,
"step": 2630
},
{
"epoch": 2.8956043956043955,
"grad_norm": 0.19825056208420425,
"learning_rate": 7.368389645252772e-08,
"loss": 0.5048,
"step": 2635
},
{
"epoch": 2.901098901098901,
"grad_norm": 0.20207333625508234,
"learning_rate": 6.61401439437348e-08,
"loss": 0.4971,
"step": 2640
},
{
"epoch": 2.9065934065934065,
"grad_norm": 0.20352522502880846,
"learning_rate": 5.9002410092262593e-08,
"loss": 0.5106,
"step": 2645
},
{
"epoch": 2.912087912087912,
"grad_norm": 0.20403786605478308,
"learning_rate": 5.227098663296404e-08,
"loss": 0.5229,
"step": 2650
},
{
"epoch": 2.9175824175824174,
"grad_norm": 0.2041806025949136,
"learning_rate": 4.594614869388947e-08,
"loss": 0.5118,
"step": 2655
},
{
"epoch": 2.9230769230769234,
"grad_norm": 0.20136601507382682,
"learning_rate": 4.002815478505007e-08,
"loss": 0.4982,
"step": 2660
},
{
"epoch": 2.928571428571429,
"grad_norm": 0.20662702533640964,
"learning_rate": 3.451724678784518e-08,
"loss": 0.5103,
"step": 2665
},
{
"epoch": 2.9340659340659343,
"grad_norm": 0.21824571202036225,
"learning_rate": 2.9413649945182475e-08,
"loss": 0.5227,
"step": 2670
},
{
"epoch": 2.9395604395604398,
"grad_norm": 0.20023301939180432,
"learning_rate": 2.47175728522675e-08,
"loss": 0.5141,
"step": 2675
},
{
"epoch": 2.9450549450549453,
"grad_norm": 0.20718821800921122,
"learning_rate": 2.0429207448078302e-08,
"loss": 0.5065,
"step": 2680
},
{
"epoch": 2.9505494505494507,
"grad_norm": 0.20188941781097472,
"learning_rate": 1.654872900752169e-08,
"loss": 0.5202,
"step": 2685
},
{
"epoch": 2.956043956043956,
"grad_norm": 0.20379828773126438,
"learning_rate": 1.3076296134271194e-08,
"loss": 0.5097,
"step": 2690
},
{
"epoch": 2.9615384615384617,
"grad_norm": 0.2043103342841997,
"learning_rate": 1.0012050754277802e-08,
"loss": 0.5127,
"step": 2695
},
{
"epoch": 2.967032967032967,
"grad_norm": 0.19873220074389136,
"learning_rate": 7.356118109977939e-09,
"loss": 0.5293,
"step": 2700
},
{
"epoch": 2.967032967032967,
"eval_loss": 0.6600582599639893,
"eval_runtime": 57.0547,
"eval_samples_per_second": 24.152,
"eval_steps_per_second": 0.105,
"step": 2700
},
{
"epoch": 2.9725274725274726,
"grad_norm": 0.20314150689373117,
"learning_rate": 5.108606755168666e-09,
"loss": 0.5146,
"step": 2705
},
{
"epoch": 2.978021978021978,
"grad_norm": 0.19470505277621175,
"learning_rate": 3.269608550571235e-09,
"loss": 0.4919,
"step": 2710
},
{
"epoch": 2.9835164835164836,
"grad_norm": 0.19203738679178015,
"learning_rate": 1.839198660079644e-09,
"loss": 0.5043,
"step": 2715
},
{
"epoch": 2.989010989010989,
"grad_norm": 0.20387845762850706,
"learning_rate": 8.174355476864293e-10,
"loss": 0.5199,
"step": 2720
},
{
"epoch": 2.9945054945054945,
"grad_norm": 0.19811095343066756,
"learning_rate": 2.0436097509235475e-10,
"loss": 0.5033,
"step": 2725
},
{
"epoch": 3.0,
"grad_norm": 0.1948663079714202,
"learning_rate": 0.0,
"loss": 0.4979,
"step": 2730
},
{
"epoch": 3.0,
"step": 2730,
"total_flos": 1.183665069490176e+16,
"train_loss": 0.5413940727492392,
"train_runtime": 63958.3461,
"train_samples_per_second": 5.459,
"train_steps_per_second": 0.043
}
],
"logging_steps": 5,
"max_steps": 2730,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.183665069490176e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}