zephyr-7b-sft-qlora / trainer_state.json
panyupj's picture
Model save
1d8b36a verified
raw
history blame
153 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 4334,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00023073373327180433,
"grad_norm": 0.686297595500946,
"learning_rate": 4.6082949308755763e-07,
"loss": 1.1177,
"step": 1
},
{
"epoch": 0.0011536686663590216,
"grad_norm": 0.7558704614639282,
"learning_rate": 2.3041474654377884e-06,
"loss": 1.1971,
"step": 5
},
{
"epoch": 0.0023073373327180432,
"grad_norm": 0.5853979587554932,
"learning_rate": 4.608294930875577e-06,
"loss": 1.1555,
"step": 10
},
{
"epoch": 0.003461005999077065,
"grad_norm": 0.6028280258178711,
"learning_rate": 6.912442396313365e-06,
"loss": 1.1553,
"step": 15
},
{
"epoch": 0.0046146746654360865,
"grad_norm": 0.5894124507904053,
"learning_rate": 9.216589861751153e-06,
"loss": 1.1362,
"step": 20
},
{
"epoch": 0.005768343331795108,
"grad_norm": 0.5509130358695984,
"learning_rate": 1.152073732718894e-05,
"loss": 1.1086,
"step": 25
},
{
"epoch": 0.00692201199815413,
"grad_norm": 0.4827423095703125,
"learning_rate": 1.382488479262673e-05,
"loss": 1.0714,
"step": 30
},
{
"epoch": 0.008075680664513151,
"grad_norm": 0.4206579923629761,
"learning_rate": 1.6129032258064517e-05,
"loss": 1.0986,
"step": 35
},
{
"epoch": 0.009229349330872173,
"grad_norm": 0.4018802344799042,
"learning_rate": 1.8433179723502307e-05,
"loss": 1.0826,
"step": 40
},
{
"epoch": 0.010383017997231195,
"grad_norm": 0.3892878293991089,
"learning_rate": 2.0737327188940094e-05,
"loss": 1.0754,
"step": 45
},
{
"epoch": 0.011536686663590217,
"grad_norm": 0.42607298493385315,
"learning_rate": 2.304147465437788e-05,
"loss": 1.0598,
"step": 50
},
{
"epoch": 0.012690355329949238,
"grad_norm": 0.4032100439071655,
"learning_rate": 2.534562211981567e-05,
"loss": 1.0398,
"step": 55
},
{
"epoch": 0.01384402399630826,
"grad_norm": 0.300424724817276,
"learning_rate": 2.764976958525346e-05,
"loss": 1.0314,
"step": 60
},
{
"epoch": 0.014997692662667282,
"grad_norm": 0.3120896816253662,
"learning_rate": 2.9953917050691244e-05,
"loss": 1.0325,
"step": 65
},
{
"epoch": 0.016151361329026302,
"grad_norm": 0.3133822977542877,
"learning_rate": 3.2258064516129034e-05,
"loss": 1.0001,
"step": 70
},
{
"epoch": 0.017305029995385326,
"grad_norm": 0.3224703371524811,
"learning_rate": 3.456221198156682e-05,
"loss": 0.9706,
"step": 75
},
{
"epoch": 0.018458698661744346,
"grad_norm": 0.3161858022212982,
"learning_rate": 3.6866359447004614e-05,
"loss": 1.0638,
"step": 80
},
{
"epoch": 0.01961236732810337,
"grad_norm": 0.3286690413951874,
"learning_rate": 3.91705069124424e-05,
"loss": 0.996,
"step": 85
},
{
"epoch": 0.02076603599446239,
"grad_norm": 0.33011969923973083,
"learning_rate": 4.147465437788019e-05,
"loss": 0.9976,
"step": 90
},
{
"epoch": 0.021919704660821413,
"grad_norm": 0.32752472162246704,
"learning_rate": 4.3778801843317974e-05,
"loss": 1.0096,
"step": 95
},
{
"epoch": 0.023073373327180433,
"grad_norm": 0.36182036995887756,
"learning_rate": 4.608294930875576e-05,
"loss": 1.0075,
"step": 100
},
{
"epoch": 0.024227041993539457,
"grad_norm": 0.3445587158203125,
"learning_rate": 4.8387096774193554e-05,
"loss": 1.002,
"step": 105
},
{
"epoch": 0.025380710659898477,
"grad_norm": 0.3711521029472351,
"learning_rate": 5.069124423963134e-05,
"loss": 0.9899,
"step": 110
},
{
"epoch": 0.0265343793262575,
"grad_norm": 0.3891366124153137,
"learning_rate": 5.2995391705069134e-05,
"loss": 1.0221,
"step": 115
},
{
"epoch": 0.02768804799261652,
"grad_norm": 0.35788699984550476,
"learning_rate": 5.529953917050692e-05,
"loss": 0.9712,
"step": 120
},
{
"epoch": 0.02884171665897554,
"grad_norm": 0.3537100553512573,
"learning_rate": 5.76036866359447e-05,
"loss": 0.9863,
"step": 125
},
{
"epoch": 0.029995385325334564,
"grad_norm": 0.3532801568508148,
"learning_rate": 5.990783410138249e-05,
"loss": 0.9795,
"step": 130
},
{
"epoch": 0.031149053991693584,
"grad_norm": 0.3615999221801758,
"learning_rate": 6.221198156682027e-05,
"loss": 1.0252,
"step": 135
},
{
"epoch": 0.032302722658052604,
"grad_norm": 0.37444955110549927,
"learning_rate": 6.451612903225807e-05,
"loss": 0.9857,
"step": 140
},
{
"epoch": 0.03345639132441163,
"grad_norm": 0.3502795696258545,
"learning_rate": 6.682027649769586e-05,
"loss": 0.9619,
"step": 145
},
{
"epoch": 0.03461005999077065,
"grad_norm": 0.40353333950042725,
"learning_rate": 6.912442396313364e-05,
"loss": 0.9964,
"step": 150
},
{
"epoch": 0.03576372865712967,
"grad_norm": 0.36437052488327026,
"learning_rate": 7.142857142857143e-05,
"loss": 1.0316,
"step": 155
},
{
"epoch": 0.03691739732348869,
"grad_norm": 0.36226072907447815,
"learning_rate": 7.373271889400923e-05,
"loss": 0.973,
"step": 160
},
{
"epoch": 0.03807106598984772,
"grad_norm": 0.33833077549934387,
"learning_rate": 7.603686635944701e-05,
"loss": 1.0132,
"step": 165
},
{
"epoch": 0.03922473465620674,
"grad_norm": 0.3641466498374939,
"learning_rate": 7.83410138248848e-05,
"loss": 0.9902,
"step": 170
},
{
"epoch": 0.04037840332256576,
"grad_norm": 0.3401365578174591,
"learning_rate": 8.064516129032258e-05,
"loss": 0.9581,
"step": 175
},
{
"epoch": 0.04153207198892478,
"grad_norm": 0.4255409836769104,
"learning_rate": 8.294930875576037e-05,
"loss": 1.0141,
"step": 180
},
{
"epoch": 0.0426857406552838,
"grad_norm": 0.34667283296585083,
"learning_rate": 8.525345622119815e-05,
"loss": 0.9797,
"step": 185
},
{
"epoch": 0.043839409321642826,
"grad_norm": 0.34790462255477905,
"learning_rate": 8.755760368663595e-05,
"loss": 1.0209,
"step": 190
},
{
"epoch": 0.044993077988001846,
"grad_norm": 0.32455047965049744,
"learning_rate": 8.986175115207374e-05,
"loss": 0.9583,
"step": 195
},
{
"epoch": 0.046146746654360866,
"grad_norm": 0.31896835565567017,
"learning_rate": 9.216589861751152e-05,
"loss": 1.0452,
"step": 200
},
{
"epoch": 0.04730041532071989,
"grad_norm": 0.35029253363609314,
"learning_rate": 9.447004608294931e-05,
"loss": 0.9895,
"step": 205
},
{
"epoch": 0.048454083987078914,
"grad_norm": 0.3429763615131378,
"learning_rate": 9.677419354838711e-05,
"loss": 1.0308,
"step": 210
},
{
"epoch": 0.049607752653437934,
"grad_norm": 0.30916911363601685,
"learning_rate": 9.907834101382489e-05,
"loss": 0.9656,
"step": 215
},
{
"epoch": 0.050761421319796954,
"grad_norm": 0.3224053680896759,
"learning_rate": 0.00010138248847926268,
"loss": 0.9742,
"step": 220
},
{
"epoch": 0.051915089986155974,
"grad_norm": 0.303576797246933,
"learning_rate": 0.00010368663594470047,
"loss": 1.0283,
"step": 225
},
{
"epoch": 0.053068758652515,
"grad_norm": 0.3039628863334656,
"learning_rate": 0.00010599078341013827,
"loss": 0.9651,
"step": 230
},
{
"epoch": 0.05422242731887402,
"grad_norm": 0.29959338903427124,
"learning_rate": 0.00010829493087557605,
"loss": 0.9649,
"step": 235
},
{
"epoch": 0.05537609598523304,
"grad_norm": 0.31773313879966736,
"learning_rate": 0.00011059907834101384,
"loss": 1.0011,
"step": 240
},
{
"epoch": 0.05652976465159206,
"grad_norm": 0.2978110909461975,
"learning_rate": 0.00011290322580645163,
"loss": 0.974,
"step": 245
},
{
"epoch": 0.05768343331795108,
"grad_norm": 0.30528807640075684,
"learning_rate": 0.0001152073732718894,
"loss": 1.0014,
"step": 250
},
{
"epoch": 0.05883710198431011,
"grad_norm": 0.2934586703777313,
"learning_rate": 0.00011751152073732718,
"loss": 1.012,
"step": 255
},
{
"epoch": 0.05999077065066913,
"grad_norm": 0.305656373500824,
"learning_rate": 0.00011981566820276497,
"loss": 0.9923,
"step": 260
},
{
"epoch": 0.06114443931702815,
"grad_norm": 0.2879476845264435,
"learning_rate": 0.00012211981566820275,
"loss": 0.9748,
"step": 265
},
{
"epoch": 0.06229810798338717,
"grad_norm": 0.2932344377040863,
"learning_rate": 0.00012442396313364055,
"loss": 0.9904,
"step": 270
},
{
"epoch": 0.06345177664974619,
"grad_norm": 0.28333616256713867,
"learning_rate": 0.00012672811059907834,
"loss": 0.9694,
"step": 275
},
{
"epoch": 0.06460544531610521,
"grad_norm": 0.3106706738471985,
"learning_rate": 0.00012903225806451613,
"loss": 0.9602,
"step": 280
},
{
"epoch": 0.06575911398246424,
"grad_norm": 0.30486252903938293,
"learning_rate": 0.00013133640552995393,
"loss": 0.997,
"step": 285
},
{
"epoch": 0.06691278264882326,
"grad_norm": 0.2930590808391571,
"learning_rate": 0.00013364055299539172,
"loss": 0.9839,
"step": 290
},
{
"epoch": 0.06806645131518228,
"grad_norm": 0.3167310655117035,
"learning_rate": 0.0001359447004608295,
"loss": 0.9784,
"step": 295
},
{
"epoch": 0.0692201199815413,
"grad_norm": 0.28749707341194153,
"learning_rate": 0.00013824884792626728,
"loss": 0.996,
"step": 300
},
{
"epoch": 0.07037378864790032,
"grad_norm": 0.33814895153045654,
"learning_rate": 0.00014055299539170507,
"loss": 1.013,
"step": 305
},
{
"epoch": 0.07152745731425934,
"grad_norm": 0.28523507714271545,
"learning_rate": 0.00014285714285714287,
"loss": 0.9904,
"step": 310
},
{
"epoch": 0.07268112598061836,
"grad_norm": 0.27578404545783997,
"learning_rate": 0.00014516129032258066,
"loss": 1.0064,
"step": 315
},
{
"epoch": 0.07383479464697738,
"grad_norm": 0.2809511721134186,
"learning_rate": 0.00014746543778801845,
"loss": 0.9846,
"step": 320
},
{
"epoch": 0.0749884633133364,
"grad_norm": 0.28782716393470764,
"learning_rate": 0.00014976958525345625,
"loss": 0.9869,
"step": 325
},
{
"epoch": 0.07614213197969544,
"grad_norm": 0.2717392146587372,
"learning_rate": 0.00015207373271889401,
"loss": 0.962,
"step": 330
},
{
"epoch": 0.07729580064605446,
"grad_norm": 0.27220791578292847,
"learning_rate": 0.0001543778801843318,
"loss": 0.9813,
"step": 335
},
{
"epoch": 0.07844946931241348,
"grad_norm": 0.3024996817111969,
"learning_rate": 0.0001566820276497696,
"loss": 0.9887,
"step": 340
},
{
"epoch": 0.0796031379787725,
"grad_norm": 0.27522554993629456,
"learning_rate": 0.0001589861751152074,
"loss": 0.9869,
"step": 345
},
{
"epoch": 0.08075680664513152,
"grad_norm": 0.27453291416168213,
"learning_rate": 0.00016129032258064516,
"loss": 0.9365,
"step": 350
},
{
"epoch": 0.08191047531149054,
"grad_norm": 0.27664312720298767,
"learning_rate": 0.00016359447004608295,
"loss": 0.9611,
"step": 355
},
{
"epoch": 0.08306414397784956,
"grad_norm": 0.2761642038822174,
"learning_rate": 0.00016589861751152075,
"loss": 0.9935,
"step": 360
},
{
"epoch": 0.08421781264420858,
"grad_norm": 0.2900699973106384,
"learning_rate": 0.00016820276497695851,
"loss": 0.9528,
"step": 365
},
{
"epoch": 0.0853714813105676,
"grad_norm": 0.2699083387851715,
"learning_rate": 0.0001705069124423963,
"loss": 0.9612,
"step": 370
},
{
"epoch": 0.08652514997692663,
"grad_norm": 0.26819881796836853,
"learning_rate": 0.0001728110599078341,
"loss": 0.9652,
"step": 375
},
{
"epoch": 0.08767881864328565,
"grad_norm": 0.2670578360557556,
"learning_rate": 0.0001751152073732719,
"loss": 0.944,
"step": 380
},
{
"epoch": 0.08883248730964467,
"grad_norm": 0.27193009853363037,
"learning_rate": 0.0001774193548387097,
"loss": 1.0175,
"step": 385
},
{
"epoch": 0.08998615597600369,
"grad_norm": 0.27275770902633667,
"learning_rate": 0.00017972350230414748,
"loss": 1.008,
"step": 390
},
{
"epoch": 0.09113982464236271,
"grad_norm": 0.2615206837654114,
"learning_rate": 0.00018202764976958527,
"loss": 1.0064,
"step": 395
},
{
"epoch": 0.09229349330872173,
"grad_norm": 0.2606607973575592,
"learning_rate": 0.00018433179723502304,
"loss": 0.9791,
"step": 400
},
{
"epoch": 0.09344716197508075,
"grad_norm": 0.257321834564209,
"learning_rate": 0.00018663594470046083,
"loss": 0.9628,
"step": 405
},
{
"epoch": 0.09460083064143977,
"grad_norm": 0.26475539803504944,
"learning_rate": 0.00018894009216589863,
"loss": 0.9771,
"step": 410
},
{
"epoch": 0.0957544993077988,
"grad_norm": 0.2725866138935089,
"learning_rate": 0.00019124423963133642,
"loss": 0.9667,
"step": 415
},
{
"epoch": 0.09690816797415783,
"grad_norm": 0.25924256443977356,
"learning_rate": 0.00019354838709677422,
"loss": 0.9865,
"step": 420
},
{
"epoch": 0.09806183664051685,
"grad_norm": 0.27873703837394714,
"learning_rate": 0.000195852534562212,
"loss": 0.9485,
"step": 425
},
{
"epoch": 0.09921550530687587,
"grad_norm": 0.26027387380599976,
"learning_rate": 0.00019815668202764977,
"loss": 1.0007,
"step": 430
},
{
"epoch": 0.10036917397323489,
"grad_norm": 0.2628461718559265,
"learning_rate": 0.00019999996755554284,
"loss": 0.971,
"step": 435
},
{
"epoch": 0.10152284263959391,
"grad_norm": 0.2701122462749481,
"learning_rate": 0.00019999883200175287,
"loss": 0.9721,
"step": 440
},
{
"epoch": 0.10267651130595293,
"grad_norm": 0.28857171535491943,
"learning_rate": 0.0001999960742461578,
"loss": 0.9881,
"step": 445
},
{
"epoch": 0.10383017997231195,
"grad_norm": 0.2599029242992401,
"learning_rate": 0.0001999916943334945,
"loss": 0.9627,
"step": 450
},
{
"epoch": 0.10498384863867097,
"grad_norm": 0.24774646759033203,
"learning_rate": 0.0001999856923348149,
"loss": 0.9885,
"step": 455
},
{
"epoch": 0.10613751730503,
"grad_norm": 0.26404890418052673,
"learning_rate": 0.00019997806834748456,
"loss": 0.9791,
"step": 460
},
{
"epoch": 0.10729118597138902,
"grad_norm": 0.25796157121658325,
"learning_rate": 0.00019996882249518144,
"loss": 0.9818,
"step": 465
},
{
"epoch": 0.10844485463774804,
"grad_norm": 0.2635892927646637,
"learning_rate": 0.0001999579549278937,
"loss": 0.9565,
"step": 470
},
{
"epoch": 0.10959852330410706,
"grad_norm": 0.2549423575401306,
"learning_rate": 0.00019994546582191718,
"loss": 0.9793,
"step": 475
},
{
"epoch": 0.11075219197046608,
"grad_norm": 0.24907591938972473,
"learning_rate": 0.00019993135537985283,
"loss": 0.9513,
"step": 480
},
{
"epoch": 0.1119058606368251,
"grad_norm": 0.2583698332309723,
"learning_rate": 0.00019991562383060317,
"loss": 0.9672,
"step": 485
},
{
"epoch": 0.11305952930318412,
"grad_norm": 0.2518406808376312,
"learning_rate": 0.00019989827142936862,
"loss": 0.984,
"step": 490
},
{
"epoch": 0.11421319796954314,
"grad_norm": 0.27680379152297974,
"learning_rate": 0.00019987929845764345,
"loss": 0.9874,
"step": 495
},
{
"epoch": 0.11536686663590216,
"grad_norm": 0.35969656705856323,
"learning_rate": 0.00019985870522321118,
"loss": 1.0057,
"step": 500
},
{
"epoch": 0.1165205353022612,
"grad_norm": 0.2667674720287323,
"learning_rate": 0.00019983649206013944,
"loss": 1.0074,
"step": 505
},
{
"epoch": 0.11767420396862022,
"grad_norm": 0.265827476978302,
"learning_rate": 0.00019981265932877488,
"loss": 0.9374,
"step": 510
},
{
"epoch": 0.11882787263497924,
"grad_norm": 0.25846490263938904,
"learning_rate": 0.00019978720741573692,
"loss": 0.9766,
"step": 515
},
{
"epoch": 0.11998154130133826,
"grad_norm": 0.24783585965633392,
"learning_rate": 0.00019976013673391182,
"loss": 0.9783,
"step": 520
},
{
"epoch": 0.12113520996769728,
"grad_norm": 0.25355273485183716,
"learning_rate": 0.00019973144772244582,
"loss": 0.9786,
"step": 525
},
{
"epoch": 0.1222888786340563,
"grad_norm": 0.2570498585700989,
"learning_rate": 0.00019970114084673796,
"loss": 0.9982,
"step": 530
},
{
"epoch": 0.12344254730041532,
"grad_norm": 0.25161927938461304,
"learning_rate": 0.00019966921659843274,
"loss": 0.9451,
"step": 535
},
{
"epoch": 0.12459621596677434,
"grad_norm": 0.2521721422672272,
"learning_rate": 0.0001996356754954119,
"loss": 0.9864,
"step": 540
},
{
"epoch": 0.12574988463313336,
"grad_norm": 0.2616737186908722,
"learning_rate": 0.00019960051808178616,
"loss": 0.9795,
"step": 545
},
{
"epoch": 0.12690355329949238,
"grad_norm": 0.25555992126464844,
"learning_rate": 0.0001995637449278864,
"loss": 0.9879,
"step": 550
},
{
"epoch": 0.1280572219658514,
"grad_norm": 0.25149908661842346,
"learning_rate": 0.0001995253566302543,
"loss": 0.952,
"step": 555
},
{
"epoch": 0.12921089063221042,
"grad_norm": 0.24779251217842102,
"learning_rate": 0.00019948535381163288,
"loss": 0.9663,
"step": 560
},
{
"epoch": 0.13036455929856944,
"grad_norm": 0.28940436244010925,
"learning_rate": 0.00019944373712095615,
"loss": 0.9649,
"step": 565
},
{
"epoch": 0.13151822796492849,
"grad_norm": 0.2656140923500061,
"learning_rate": 0.00019940050723333866,
"loss": 0.9729,
"step": 570
},
{
"epoch": 0.1326718966312875,
"grad_norm": 0.26921287178993225,
"learning_rate": 0.00019935566485006464,
"loss": 0.9812,
"step": 575
},
{
"epoch": 0.13382556529764653,
"grad_norm": 0.2550528943538666,
"learning_rate": 0.0001993092106985765,
"loss": 0.9711,
"step": 580
},
{
"epoch": 0.13497923396400555,
"grad_norm": 0.26933753490448,
"learning_rate": 0.0001992611455324632,
"loss": 0.9782,
"step": 585
},
{
"epoch": 0.13613290263036457,
"grad_norm": 0.25276893377304077,
"learning_rate": 0.0001992114701314478,
"loss": 0.9438,
"step": 590
},
{
"epoch": 0.1372865712967236,
"grad_norm": 0.2518392503261566,
"learning_rate": 0.00019916018530137495,
"loss": 0.9835,
"step": 595
},
{
"epoch": 0.1384402399630826,
"grad_norm": 0.2546631693840027,
"learning_rate": 0.00019910729187419781,
"loss": 1.0105,
"step": 600
},
{
"epoch": 0.13959390862944163,
"grad_norm": 0.26017695665359497,
"learning_rate": 0.00019905279070796454,
"loss": 0.9554,
"step": 605
},
{
"epoch": 0.14074757729580065,
"grad_norm": 0.2573865056037903,
"learning_rate": 0.0001989966826868044,
"loss": 0.9939,
"step": 610
},
{
"epoch": 0.14190124596215967,
"grad_norm": 0.25712475180625916,
"learning_rate": 0.00019893896872091329,
"loss": 0.9359,
"step": 615
},
{
"epoch": 0.1430549146285187,
"grad_norm": 0.2623005509376526,
"learning_rate": 0.00019887964974653918,
"loss": 0.9486,
"step": 620
},
{
"epoch": 0.1442085832948777,
"grad_norm": 0.2532264292240143,
"learning_rate": 0.00019881872672596683,
"loss": 0.9803,
"step": 625
},
{
"epoch": 0.14536225196123673,
"grad_norm": 0.2683295011520386,
"learning_rate": 0.00019875620064750202,
"loss": 0.9611,
"step": 630
},
{
"epoch": 0.14651592062759575,
"grad_norm": 0.26294204592704773,
"learning_rate": 0.00019869207252545584,
"loss": 0.9749,
"step": 635
},
{
"epoch": 0.14766958929395477,
"grad_norm": 0.24312348663806915,
"learning_rate": 0.00019862634340012795,
"loss": 0.9444,
"step": 640
},
{
"epoch": 0.1488232579603138,
"grad_norm": 0.25715410709381104,
"learning_rate": 0.00019855901433778991,
"loss": 0.9491,
"step": 645
},
{
"epoch": 0.1499769266266728,
"grad_norm": 0.2563926875591278,
"learning_rate": 0.00019849008643066772,
"loss": 0.9484,
"step": 650
},
{
"epoch": 0.15113059529303183,
"grad_norm": 0.2505110800266266,
"learning_rate": 0.0001984195607969242,
"loss": 0.9911,
"step": 655
},
{
"epoch": 0.15228426395939088,
"grad_norm": 0.2632611095905304,
"learning_rate": 0.0001983474385806408,
"loss": 0.9644,
"step": 660
},
{
"epoch": 0.1534379326257499,
"grad_norm": 0.25112318992614746,
"learning_rate": 0.00019827372095179908,
"loss": 0.9766,
"step": 665
},
{
"epoch": 0.15459160129210892,
"grad_norm": 0.2595708668231964,
"learning_rate": 0.00019819840910626174,
"loss": 0.9931,
"step": 670
},
{
"epoch": 0.15574526995846794,
"grad_norm": 0.2515600621700287,
"learning_rate": 0.00019812150426575315,
"loss": 0.9581,
"step": 675
},
{
"epoch": 0.15689893862482696,
"grad_norm": 0.2485395222902298,
"learning_rate": 0.00019804300767783958,
"loss": 0.9339,
"step": 680
},
{
"epoch": 0.15805260729118598,
"grad_norm": 0.25041699409484863,
"learning_rate": 0.00019796292061590899,
"loss": 0.9478,
"step": 685
},
{
"epoch": 0.159206275957545,
"grad_norm": 0.26130861043930054,
"learning_rate": 0.0001978812443791503,
"loss": 0.975,
"step": 690
},
{
"epoch": 0.16035994462390402,
"grad_norm": 0.2546372711658478,
"learning_rate": 0.00019779798029253241,
"loss": 0.9673,
"step": 695
},
{
"epoch": 0.16151361329026304,
"grad_norm": 0.2580021023750305,
"learning_rate": 0.00019771312970678258,
"loss": 0.9921,
"step": 700
},
{
"epoch": 0.16266728195662206,
"grad_norm": 0.24550603330135345,
"learning_rate": 0.00019762669399836462,
"loss": 0.9568,
"step": 705
},
{
"epoch": 0.16382095062298108,
"grad_norm": 0.25711777806282043,
"learning_rate": 0.0001975386745694565,
"loss": 0.9671,
"step": 710
},
{
"epoch": 0.1649746192893401,
"grad_norm": 0.2552441656589508,
"learning_rate": 0.00019744907284792776,
"loss": 0.9796,
"step": 715
},
{
"epoch": 0.16612828795569912,
"grad_norm": 0.24158374965190887,
"learning_rate": 0.00019735789028731604,
"loss": 0.9841,
"step": 720
},
{
"epoch": 0.16728195662205814,
"grad_norm": 0.25179579854011536,
"learning_rate": 0.00019726512836680378,
"loss": 0.9817,
"step": 725
},
{
"epoch": 0.16843562528841716,
"grad_norm": 0.27372288703918457,
"learning_rate": 0.0001971707885911941,
"loss": 0.9654,
"step": 730
},
{
"epoch": 0.16958929395477618,
"grad_norm": 0.28059613704681396,
"learning_rate": 0.00019707487249088641,
"loss": 0.9669,
"step": 735
},
{
"epoch": 0.1707429626211352,
"grad_norm": 0.27208980917930603,
"learning_rate": 0.00019697738162185161,
"loss": 1.0043,
"step": 740
},
{
"epoch": 0.17189663128749424,
"grad_norm": 0.25409960746765137,
"learning_rate": 0.0001968783175656068,
"loss": 0.9751,
"step": 745
},
{
"epoch": 0.17305029995385326,
"grad_norm": 0.2589627504348755,
"learning_rate": 0.00019677768192918971,
"loss": 0.9733,
"step": 750
},
{
"epoch": 0.17420396862021229,
"grad_norm": 0.2695959210395813,
"learning_rate": 0.00019667547634513247,
"loss": 0.955,
"step": 755
},
{
"epoch": 0.1753576372865713,
"grad_norm": 0.25583890080451965,
"learning_rate": 0.00019657170247143525,
"loss": 0.955,
"step": 760
},
{
"epoch": 0.17651130595293033,
"grad_norm": 0.24999220669269562,
"learning_rate": 0.0001964663619915394,
"loss": 0.9715,
"step": 765
},
{
"epoch": 0.17766497461928935,
"grad_norm": 0.2531087100505829,
"learning_rate": 0.00019635945661430006,
"loss": 0.9917,
"step": 770
},
{
"epoch": 0.17881864328564837,
"grad_norm": 0.2580994665622711,
"learning_rate": 0.0001962509880739584,
"loss": 0.9876,
"step": 775
},
{
"epoch": 0.17997231195200739,
"grad_norm": 0.2500920593738556,
"learning_rate": 0.00019614095813011364,
"loss": 0.9964,
"step": 780
},
{
"epoch": 0.1811259806183664,
"grad_norm": 0.2500130534172058,
"learning_rate": 0.0001960293685676943,
"loss": 0.9536,
"step": 785
},
{
"epoch": 0.18227964928472543,
"grad_norm": 0.25024518370628357,
"learning_rate": 0.0001959162211969295,
"loss": 0.961,
"step": 790
},
{
"epoch": 0.18343331795108445,
"grad_norm": 0.25615155696868896,
"learning_rate": 0.00019580151785331934,
"loss": 0.9795,
"step": 795
},
{
"epoch": 0.18458698661744347,
"grad_norm": 0.2746971547603607,
"learning_rate": 0.0001956852603976052,
"loss": 0.99,
"step": 800
},
{
"epoch": 0.18574065528380249,
"grad_norm": 0.2582091987133026,
"learning_rate": 0.00019556745071573975,
"loss": 0.9599,
"step": 805
},
{
"epoch": 0.1868943239501615,
"grad_norm": 0.25607603788375854,
"learning_rate": 0.00019544809071885604,
"loss": 0.9739,
"step": 810
},
{
"epoch": 0.18804799261652053,
"grad_norm": 0.25475266575813293,
"learning_rate": 0.00019532718234323672,
"loss": 0.9638,
"step": 815
},
{
"epoch": 0.18920166128287955,
"grad_norm": 0.2636029124259949,
"learning_rate": 0.00019520472755028256,
"loss": 1.0138,
"step": 820
},
{
"epoch": 0.19035532994923857,
"grad_norm": 0.25477829575538635,
"learning_rate": 0.00019508072832648062,
"loss": 0.928,
"step": 825
},
{
"epoch": 0.1915089986155976,
"grad_norm": 0.2600194215774536,
"learning_rate": 0.00019495518668337201,
"loss": 0.98,
"step": 830
},
{
"epoch": 0.19266266728195663,
"grad_norm": 0.2581029534339905,
"learning_rate": 0.00019482810465751938,
"loss": 0.9466,
"step": 835
},
{
"epoch": 0.19381633594831565,
"grad_norm": 0.26259845495224,
"learning_rate": 0.0001946994843104737,
"loss": 0.9702,
"step": 840
},
{
"epoch": 0.19497000461467467,
"grad_norm": 0.2556231617927551,
"learning_rate": 0.00019456932772874091,
"loss": 0.9752,
"step": 845
},
{
"epoch": 0.1961236732810337,
"grad_norm": 0.25811153650283813,
"learning_rate": 0.00019443763702374812,
"loss": 0.9616,
"step": 850
},
{
"epoch": 0.19727734194739271,
"grad_norm": 0.25802063941955566,
"learning_rate": 0.00019430441433180925,
"loss": 0.9736,
"step": 855
},
{
"epoch": 0.19843101061375173,
"grad_norm": 0.27993249893188477,
"learning_rate": 0.00019416966181409046,
"loss": 0.9527,
"step": 860
},
{
"epoch": 0.19958467928011075,
"grad_norm": 0.2626069188117981,
"learning_rate": 0.00019403338165657508,
"loss": 0.955,
"step": 865
},
{
"epoch": 0.20073834794646978,
"grad_norm": 0.2545926570892334,
"learning_rate": 0.00019389557607002805,
"loss": 0.9896,
"step": 870
},
{
"epoch": 0.2018920166128288,
"grad_norm": 0.2604697346687317,
"learning_rate": 0.0001937562472899603,
"loss": 0.9769,
"step": 875
},
{
"epoch": 0.20304568527918782,
"grad_norm": 0.2628583312034607,
"learning_rate": 0.0001936153975765921,
"loss": 0.9392,
"step": 880
},
{
"epoch": 0.20419935394554684,
"grad_norm": 0.2587975859642029,
"learning_rate": 0.00019347302921481681,
"loss": 0.9933,
"step": 885
},
{
"epoch": 0.20535302261190586,
"grad_norm": 0.27935993671417236,
"learning_rate": 0.00019332914451416347,
"loss": 0.9232,
"step": 890
},
{
"epoch": 0.20650669127826488,
"grad_norm": 0.2539782226085663,
"learning_rate": 0.00019318374580875962,
"loss": 0.9781,
"step": 895
},
{
"epoch": 0.2076603599446239,
"grad_norm": 0.27198004722595215,
"learning_rate": 0.00019303683545729322,
"loss": 0.9484,
"step": 900
},
{
"epoch": 0.20881402861098292,
"grad_norm": 0.25046050548553467,
"learning_rate": 0.00019288841584297445,
"loss": 0.9693,
"step": 905
},
{
"epoch": 0.20996769727734194,
"grad_norm": 0.2635994851589203,
"learning_rate": 0.0001927384893734971,
"loss": 0.9412,
"step": 910
},
{
"epoch": 0.21112136594370096,
"grad_norm": 0.26054415106773376,
"learning_rate": 0.0001925870584809995,
"loss": 0.9485,
"step": 915
},
{
"epoch": 0.21227503461006,
"grad_norm": 0.25425177812576294,
"learning_rate": 0.00019243412562202497,
"loss": 0.9841,
"step": 920
},
{
"epoch": 0.21342870327641902,
"grad_norm": 0.25780272483825684,
"learning_rate": 0.00019227969327748211,
"loss": 0.9502,
"step": 925
},
{
"epoch": 0.21458237194277804,
"grad_norm": 0.24851177632808685,
"learning_rate": 0.00019212376395260448,
"loss": 0.9875,
"step": 930
},
{
"epoch": 0.21573604060913706,
"grad_norm": 0.25959885120391846,
"learning_rate": 0.00019196634017690993,
"loss": 0.9734,
"step": 935
},
{
"epoch": 0.21688970927549608,
"grad_norm": 0.2571864426136017,
"learning_rate": 0.00019180742450415964,
"loss": 0.9636,
"step": 940
},
{
"epoch": 0.2180433779418551,
"grad_norm": 0.2478172332048416,
"learning_rate": 0.00019164701951231657,
"loss": 0.9718,
"step": 945
},
{
"epoch": 0.21919704660821412,
"grad_norm": 0.27649134397506714,
"learning_rate": 0.00019148512780350384,
"loss": 0.9737,
"step": 950
},
{
"epoch": 0.22035071527457314,
"grad_norm": 0.25620779395103455,
"learning_rate": 0.00019132175200396235,
"loss": 0.9431,
"step": 955
},
{
"epoch": 0.22150438394093216,
"grad_norm": 0.2638430595397949,
"learning_rate": 0.00019115689476400816,
"loss": 0.984,
"step": 960
},
{
"epoch": 0.22265805260729118,
"grad_norm": 0.25410696864128113,
"learning_rate": 0.00019099055875798973,
"loss": 0.9703,
"step": 965
},
{
"epoch": 0.2238117212736502,
"grad_norm": 0.25862041115760803,
"learning_rate": 0.00019082274668424422,
"loss": 0.9509,
"step": 970
},
{
"epoch": 0.22496538994000922,
"grad_norm": 0.24762850999832153,
"learning_rate": 0.0001906534612650539,
"loss": 0.9424,
"step": 975
},
{
"epoch": 0.22611905860636825,
"grad_norm": 0.2532382309436798,
"learning_rate": 0.00019048270524660196,
"loss": 1.0109,
"step": 980
},
{
"epoch": 0.22727272727272727,
"grad_norm": 0.296761155128479,
"learning_rate": 0.000190310481398928,
"loss": 0.9461,
"step": 985
},
{
"epoch": 0.22842639593908629,
"grad_norm": 0.2572844624519348,
"learning_rate": 0.00019013679251588303,
"loss": 0.9685,
"step": 990
},
{
"epoch": 0.2295800646054453,
"grad_norm": 0.2736752927303314,
"learning_rate": 0.00018996164141508412,
"loss": 0.9788,
"step": 995
},
{
"epoch": 0.23073373327180433,
"grad_norm": 0.26040878891944885,
"learning_rate": 0.00018978503093786882,
"loss": 0.9669,
"step": 1000
},
{
"epoch": 0.23188740193816337,
"grad_norm": 0.26469501852989197,
"learning_rate": 0.00018960696394924886,
"loss": 1.0047,
"step": 1005
},
{
"epoch": 0.2330410706045224,
"grad_norm": 0.25627848505973816,
"learning_rate": 0.00018942744333786397,
"loss": 0.9911,
"step": 1010
},
{
"epoch": 0.2341947392708814,
"grad_norm": 0.2675924003124237,
"learning_rate": 0.00018924647201593473,
"loss": 0.9934,
"step": 1015
},
{
"epoch": 0.23534840793724043,
"grad_norm": 0.2544163763523102,
"learning_rate": 0.00018906405291921547,
"loss": 0.9674,
"step": 1020
},
{
"epoch": 0.23650207660359945,
"grad_norm": 0.2576877772808075,
"learning_rate": 0.00018888018900694667,
"loss": 0.9345,
"step": 1025
},
{
"epoch": 0.23765574526995847,
"grad_norm": 0.35056746006011963,
"learning_rate": 0.00018869488326180679,
"loss": 0.9842,
"step": 1030
},
{
"epoch": 0.2388094139363175,
"grad_norm": 0.25830191373825073,
"learning_rate": 0.00018850813868986416,
"loss": 0.9229,
"step": 1035
},
{
"epoch": 0.23996308260267651,
"grad_norm": 0.25119414925575256,
"learning_rate": 0.000188319958320528,
"loss": 0.9594,
"step": 1040
},
{
"epoch": 0.24111675126903553,
"grad_norm": 0.2568560242652893,
"learning_rate": 0.0001881303452064992,
"loss": 0.9645,
"step": 1045
},
{
"epoch": 0.24227041993539455,
"grad_norm": 0.2648126482963562,
"learning_rate": 0.0001879393024237212,
"loss": 0.9627,
"step": 1050
},
{
"epoch": 0.24342408860175357,
"grad_norm": 0.2568890154361725,
"learning_rate": 0.00018774683307132953,
"loss": 0.9595,
"step": 1055
},
{
"epoch": 0.2445777572681126,
"grad_norm": 0.26714202761650085,
"learning_rate": 0.00018755294027160204,
"loss": 0.9474,
"step": 1060
},
{
"epoch": 0.24573142593447161,
"grad_norm": 0.2748320996761322,
"learning_rate": 0.00018735762716990797,
"loss": 0.9474,
"step": 1065
},
{
"epoch": 0.24688509460083063,
"grad_norm": 0.28429654240608215,
"learning_rate": 0.00018716089693465696,
"loss": 0.99,
"step": 1070
},
{
"epoch": 0.24803876326718965,
"grad_norm": 0.2595331072807312,
"learning_rate": 0.0001869627527572477,
"loss": 0.9592,
"step": 1075
},
{
"epoch": 0.24919243193354867,
"grad_norm": 0.2595003843307495,
"learning_rate": 0.00018676319785201616,
"loss": 0.9882,
"step": 1080
},
{
"epoch": 0.2503461005999077,
"grad_norm": 0.3043324649333954,
"learning_rate": 0.00018656223545618345,
"loss": 0.9573,
"step": 1085
},
{
"epoch": 0.2514997692662667,
"grad_norm": 0.2606871724128723,
"learning_rate": 0.00018635986882980325,
"loss": 0.9808,
"step": 1090
},
{
"epoch": 0.25265343793262574,
"grad_norm": 0.26122790575027466,
"learning_rate": 0.00018615610125570902,
"loss": 0.9523,
"step": 1095
},
{
"epoch": 0.25380710659898476,
"grad_norm": 0.2673528492450714,
"learning_rate": 0.00018595093603946053,
"loss": 0.9495,
"step": 1100
},
{
"epoch": 0.2549607752653438,
"grad_norm": 0.25935789942741394,
"learning_rate": 0.00018574437650929054,
"loss": 0.9026,
"step": 1105
},
{
"epoch": 0.2561144439317028,
"grad_norm": 0.28227198123931885,
"learning_rate": 0.00018553642601605068,
"loss": 0.9456,
"step": 1110
},
{
"epoch": 0.2572681125980618,
"grad_norm": 0.25894030928611755,
"learning_rate": 0.00018532708793315693,
"loss": 0.987,
"step": 1115
},
{
"epoch": 0.25842178126442084,
"grad_norm": 0.2698533535003662,
"learning_rate": 0.00018511636565653511,
"loss": 0.9778,
"step": 1120
},
{
"epoch": 0.25957544993077986,
"grad_norm": 0.25481945276260376,
"learning_rate": 0.00018490426260456578,
"loss": 0.978,
"step": 1125
},
{
"epoch": 0.2607291185971389,
"grad_norm": 0.28665271401405334,
"learning_rate": 0.0001846907822180286,
"loss": 0.9764,
"step": 1130
},
{
"epoch": 0.2618827872634979,
"grad_norm": 0.2624281942844391,
"learning_rate": 0.00018447592796004675,
"loss": 0.9534,
"step": 1135
},
{
"epoch": 0.26303645592985697,
"grad_norm": 0.2616020143032074,
"learning_rate": 0.00018425970331603056,
"loss": 0.9692,
"step": 1140
},
{
"epoch": 0.264190124596216,
"grad_norm": 0.2609774172306061,
"learning_rate": 0.00018404211179362114,
"loss": 0.9681,
"step": 1145
},
{
"epoch": 0.265343793262575,
"grad_norm": 0.2684124708175659,
"learning_rate": 0.00018382315692263323,
"loss": 1.005,
"step": 1150
},
{
"epoch": 0.26649746192893403,
"grad_norm": 0.2579489052295685,
"learning_rate": 0.0001836028422549983,
"loss": 1.0064,
"step": 1155
},
{
"epoch": 0.26765113059529305,
"grad_norm": 0.26837727427482605,
"learning_rate": 0.00018338117136470648,
"loss": 0.9873,
"step": 1160
},
{
"epoch": 0.26880479926165207,
"grad_norm": 0.27158886194229126,
"learning_rate": 0.00018315814784774901,
"loss": 0.9775,
"step": 1165
},
{
"epoch": 0.2699584679280111,
"grad_norm": 0.28705140948295593,
"learning_rate": 0.00018293377532205968,
"loss": 0.9876,
"step": 1170
},
{
"epoch": 0.2711121365943701,
"grad_norm": 0.2695893347263336,
"learning_rate": 0.00018270805742745617,
"loss": 0.9651,
"step": 1175
},
{
"epoch": 0.27226580526072913,
"grad_norm": 0.2642047703266144,
"learning_rate": 0.000182480997825581,
"loss": 0.9849,
"step": 1180
},
{
"epoch": 0.27341947392708815,
"grad_norm": 0.24997040629386902,
"learning_rate": 0.00018225260019984215,
"loss": 0.9574,
"step": 1185
},
{
"epoch": 0.2745731425934472,
"grad_norm": 0.25589871406555176,
"learning_rate": 0.0001820228682553533,
"loss": 0.9311,
"step": 1190
},
{
"epoch": 0.2757268112598062,
"grad_norm": 0.2490740865468979,
"learning_rate": 0.00018179180571887372,
"loss": 0.9471,
"step": 1195
},
{
"epoch": 0.2768804799261652,
"grad_norm": 0.24929802119731903,
"learning_rate": 0.00018155941633874787,
"loss": 0.9551,
"step": 1200
},
{
"epoch": 0.27803414859252423,
"grad_norm": 0.2558966875076294,
"learning_rate": 0.00018132570388484443,
"loss": 0.9283,
"step": 1205
},
{
"epoch": 0.27918781725888325,
"grad_norm": 0.26152223348617554,
"learning_rate": 0.00018109067214849538,
"loss": 0.9457,
"step": 1210
},
{
"epoch": 0.2803414859252423,
"grad_norm": 0.2626384198665619,
"learning_rate": 0.0001808543249424343,
"loss": 0.9678,
"step": 1215
},
{
"epoch": 0.2814951545916013,
"grad_norm": 0.25507864356040955,
"learning_rate": 0.00018061666610073464,
"loss": 0.9639,
"step": 1220
},
{
"epoch": 0.2826488232579603,
"grad_norm": 0.27339187264442444,
"learning_rate": 0.00018037769947874742,
"loss": 0.9475,
"step": 1225
},
{
"epoch": 0.28380249192431933,
"grad_norm": 0.26297956705093384,
"learning_rate": 0.00018013742895303883,
"loss": 0.9415,
"step": 1230
},
{
"epoch": 0.28495616059067835,
"grad_norm": 0.2605285942554474,
"learning_rate": 0.00017989585842132712,
"loss": 0.9507,
"step": 1235
},
{
"epoch": 0.2861098292570374,
"grad_norm": 0.2680794596672058,
"learning_rate": 0.00017965299180241963,
"loss": 0.9928,
"step": 1240
},
{
"epoch": 0.2872634979233964,
"grad_norm": 0.2609361410140991,
"learning_rate": 0.00017940883303614905,
"loss": 1.0034,
"step": 1245
},
{
"epoch": 0.2884171665897554,
"grad_norm": 0.26120924949645996,
"learning_rate": 0.0001791633860833096,
"loss": 0.9703,
"step": 1250
},
{
"epoch": 0.28957083525611443,
"grad_norm": 0.259295791387558,
"learning_rate": 0.0001789166549255926,
"loss": 0.9717,
"step": 1255
},
{
"epoch": 0.29072450392247345,
"grad_norm": 0.25710657238960266,
"learning_rate": 0.00017866864356552213,
"loss": 0.9773,
"step": 1260
},
{
"epoch": 0.2918781725888325,
"grad_norm": 0.25823676586151123,
"learning_rate": 0.00017841935602638996,
"loss": 0.9771,
"step": 1265
},
{
"epoch": 0.2930318412551915,
"grad_norm": 0.2612430155277252,
"learning_rate": 0.00017816879635219028,
"loss": 0.9257,
"step": 1270
},
{
"epoch": 0.2941855099215505,
"grad_norm": 0.2545841932296753,
"learning_rate": 0.0001779169686075541,
"loss": 0.931,
"step": 1275
},
{
"epoch": 0.29533917858790953,
"grad_norm": 0.2565891146659851,
"learning_rate": 0.0001776638768776834,
"loss": 0.9442,
"step": 1280
},
{
"epoch": 0.29649284725426855,
"grad_norm": 0.2589223384857178,
"learning_rate": 0.00017740952526828466,
"loss": 0.9319,
"step": 1285
},
{
"epoch": 0.2976465159206276,
"grad_norm": 0.26724082231521606,
"learning_rate": 0.00017715391790550252,
"loss": 0.9854,
"step": 1290
},
{
"epoch": 0.2988001845869866,
"grad_norm": 0.2645433247089386,
"learning_rate": 0.00017689705893585272,
"loss": 0.9683,
"step": 1295
},
{
"epoch": 0.2999538532533456,
"grad_norm": 0.2574128210544586,
"learning_rate": 0.0001766389525261547,
"loss": 0.987,
"step": 1300
},
{
"epoch": 0.30110752191970463,
"grad_norm": 0.2574234902858734,
"learning_rate": 0.00017637960286346425,
"loss": 0.9991,
"step": 1305
},
{
"epoch": 0.30226119058606365,
"grad_norm": 0.255074143409729,
"learning_rate": 0.00017611901415500535,
"loss": 1.0042,
"step": 1310
},
{
"epoch": 0.30341485925242273,
"grad_norm": 0.26597529649734497,
"learning_rate": 0.00017585719062810214,
"loss": 0.9865,
"step": 1315
},
{
"epoch": 0.30456852791878175,
"grad_norm": 0.2626805901527405,
"learning_rate": 0.00017559413653011024,
"loss": 0.9751,
"step": 1320
},
{
"epoch": 0.30572219658514077,
"grad_norm": 0.24981103837490082,
"learning_rate": 0.0001753298561283478,
"loss": 0.9628,
"step": 1325
},
{
"epoch": 0.3068758652514998,
"grad_norm": 0.25727543234825134,
"learning_rate": 0.00017506435371002633,
"loss": 0.9427,
"step": 1330
},
{
"epoch": 0.3080295339178588,
"grad_norm": 0.27170515060424805,
"learning_rate": 0.00017479763358218119,
"loss": 0.9777,
"step": 1335
},
{
"epoch": 0.30918320258421783,
"grad_norm": 0.27030548453330994,
"learning_rate": 0.0001745297000716016,
"loss": 0.9678,
"step": 1340
},
{
"epoch": 0.31033687125057685,
"grad_norm": 0.27044013142585754,
"learning_rate": 0.0001742605575247606,
"loss": 1.0005,
"step": 1345
},
{
"epoch": 0.31149053991693587,
"grad_norm": 0.2567753195762634,
"learning_rate": 0.00017399021030774442,
"loss": 0.9379,
"step": 1350
},
{
"epoch": 0.3126442085832949,
"grad_norm": 0.2584477961063385,
"learning_rate": 0.00017371866280618177,
"loss": 0.9599,
"step": 1355
},
{
"epoch": 0.3137978772496539,
"grad_norm": 0.26751193404197693,
"learning_rate": 0.0001734459194251725,
"loss": 0.9778,
"step": 1360
},
{
"epoch": 0.31495154591601293,
"grad_norm": 0.26162561774253845,
"learning_rate": 0.00017317198458921638,
"loss": 0.9809,
"step": 1365
},
{
"epoch": 0.31610521458237195,
"grad_norm": 0.26419851183891296,
"learning_rate": 0.00017289686274214118,
"loss": 0.9779,
"step": 1370
},
{
"epoch": 0.31725888324873097,
"grad_norm": 0.25940561294555664,
"learning_rate": 0.00017262055834703054,
"loss": 0.9899,
"step": 1375
},
{
"epoch": 0.31841255191509,
"grad_norm": 0.2542431652545929,
"learning_rate": 0.00017234307588615176,
"loss": 0.9748,
"step": 1380
},
{
"epoch": 0.319566220581449,
"grad_norm": 0.25366565585136414,
"learning_rate": 0.0001720644198608829,
"loss": 0.9828,
"step": 1385
},
{
"epoch": 0.32071988924780803,
"grad_norm": 0.2604008913040161,
"learning_rate": 0.00017178459479163976,
"loss": 0.9697,
"step": 1390
},
{
"epoch": 0.32187355791416705,
"grad_norm": 0.2703900933265686,
"learning_rate": 0.0001715036052178028,
"loss": 0.986,
"step": 1395
},
{
"epoch": 0.3230272265805261,
"grad_norm": 0.25708577036857605,
"learning_rate": 0.0001712214556976431,
"loss": 0.9457,
"step": 1400
},
{
"epoch": 0.3241808952468851,
"grad_norm": 0.26468509435653687,
"learning_rate": 0.00017093815080824876,
"loss": 0.9599,
"step": 1405
},
{
"epoch": 0.3253345639132441,
"grad_norm": 0.2600449025630951,
"learning_rate": 0.00017065369514545053,
"loss": 0.941,
"step": 1410
},
{
"epoch": 0.32648823257960313,
"grad_norm": 0.2549765110015869,
"learning_rate": 0.00017036809332374712,
"loss": 0.9735,
"step": 1415
},
{
"epoch": 0.32764190124596215,
"grad_norm": 0.2710689902305603,
"learning_rate": 0.00017008134997623065,
"loss": 0.9949,
"step": 1420
},
{
"epoch": 0.3287955699123212,
"grad_norm": 0.2554212808609009,
"learning_rate": 0.00016979346975451112,
"loss": 0.9741,
"step": 1425
},
{
"epoch": 0.3299492385786802,
"grad_norm": 0.2632601261138916,
"learning_rate": 0.00016950445732864127,
"loss": 0.9951,
"step": 1430
},
{
"epoch": 0.3311029072450392,
"grad_norm": 0.2675810754299164,
"learning_rate": 0.0001692143173870407,
"loss": 0.9544,
"step": 1435
},
{
"epoch": 0.33225657591139823,
"grad_norm": 0.26224300265312195,
"learning_rate": 0.00016892305463641965,
"loss": 0.9297,
"step": 1440
},
{
"epoch": 0.33341024457775725,
"grad_norm": 0.2826540470123291,
"learning_rate": 0.00016863067380170298,
"loss": 0.9588,
"step": 1445
},
{
"epoch": 0.3345639132441163,
"grad_norm": 0.2714956998825073,
"learning_rate": 0.00016833717962595326,
"loss": 0.9682,
"step": 1450
},
{
"epoch": 0.3357175819104753,
"grad_norm": 0.24968619644641876,
"learning_rate": 0.00016804257687029389,
"loss": 1.0191,
"step": 1455
},
{
"epoch": 0.3368712505768343,
"grad_norm": 0.26332545280456543,
"learning_rate": 0.00016774687031383188,
"loss": 0.957,
"step": 1460
},
{
"epoch": 0.33802491924319333,
"grad_norm": 0.25739216804504395,
"learning_rate": 0.00016745006475358046,
"loss": 0.9747,
"step": 1465
},
{
"epoch": 0.33917858790955235,
"grad_norm": 0.26346343755722046,
"learning_rate": 0.00016715216500438093,
"loss": 0.9414,
"step": 1470
},
{
"epoch": 0.3403322565759114,
"grad_norm": 0.25862592458724976,
"learning_rate": 0.0001668531758988249,
"loss": 0.962,
"step": 1475
},
{
"epoch": 0.3414859252422704,
"grad_norm": 0.25549453496932983,
"learning_rate": 0.00016655310228717564,
"loss": 0.9575,
"step": 1480
},
{
"epoch": 0.3426395939086294,
"grad_norm": 0.26341837644577026,
"learning_rate": 0.0001662519490372896,
"loss": 0.9546,
"step": 1485
},
{
"epoch": 0.3437932625749885,
"grad_norm": 0.279466450214386,
"learning_rate": 0.00016594972103453726,
"loss": 0.9755,
"step": 1490
},
{
"epoch": 0.3449469312413475,
"grad_norm": 0.2663206160068512,
"learning_rate": 0.00016564642318172402,
"loss": 0.9683,
"step": 1495
},
{
"epoch": 0.34610059990770653,
"grad_norm": 0.2633247971534729,
"learning_rate": 0.00016534206039901057,
"loss": 0.9425,
"step": 1500
},
{
"epoch": 0.34725426857406555,
"grad_norm": 0.26610881090164185,
"learning_rate": 0.00016503663762383312,
"loss": 0.9571,
"step": 1505
},
{
"epoch": 0.34840793724042457,
"grad_norm": 0.25422990322113037,
"learning_rate": 0.00016473015981082338,
"loss": 0.965,
"step": 1510
},
{
"epoch": 0.3495616059067836,
"grad_norm": 0.26195505261421204,
"learning_rate": 0.000164422631931728,
"loss": 0.9811,
"step": 1515
},
{
"epoch": 0.3507152745731426,
"grad_norm": 0.2687501609325409,
"learning_rate": 0.00016411405897532802,
"loss": 0.9657,
"step": 1520
},
{
"epoch": 0.35186894323950163,
"grad_norm": 0.27142685651779175,
"learning_rate": 0.000163804445947358,
"loss": 0.9656,
"step": 1525
},
{
"epoch": 0.35302261190586065,
"grad_norm": 0.2657437026500702,
"learning_rate": 0.00016349379787042477,
"loss": 0.9561,
"step": 1530
},
{
"epoch": 0.35417628057221967,
"grad_norm": 0.266258180141449,
"learning_rate": 0.00016318211978392589,
"loss": 0.9611,
"step": 1535
},
{
"epoch": 0.3553299492385787,
"grad_norm": 0.2663586735725403,
"learning_rate": 0.00016286941674396787,
"loss": 0.9368,
"step": 1540
},
{
"epoch": 0.3564836179049377,
"grad_norm": 0.2677685618400574,
"learning_rate": 0.00016255569382328443,
"loss": 0.9867,
"step": 1545
},
{
"epoch": 0.35763728657129673,
"grad_norm": 0.25513461232185364,
"learning_rate": 0.00016224095611115384,
"loss": 0.9874,
"step": 1550
},
{
"epoch": 0.35879095523765575,
"grad_norm": 0.24931229650974274,
"learning_rate": 0.00016192520871331661,
"loss": 0.9173,
"step": 1555
},
{
"epoch": 0.35994462390401477,
"grad_norm": 0.25578710436820984,
"learning_rate": 0.00016160845675189254,
"loss": 0.9848,
"step": 1560
},
{
"epoch": 0.3610982925703738,
"grad_norm": 0.26572901010513306,
"learning_rate": 0.00016129070536529766,
"loss": 0.9756,
"step": 1565
},
{
"epoch": 0.3622519612367328,
"grad_norm": 0.28126901388168335,
"learning_rate": 0.00016097195970816094,
"loss": 0.9627,
"step": 1570
},
{
"epoch": 0.36340562990309183,
"grad_norm": 0.27309858798980713,
"learning_rate": 0.00016065222495124056,
"loss": 0.9393,
"step": 1575
},
{
"epoch": 0.36455929856945085,
"grad_norm": 0.2919960021972656,
"learning_rate": 0.00016033150628134011,
"loss": 0.9475,
"step": 1580
},
{
"epoch": 0.36571296723580987,
"grad_norm": 0.2607177495956421,
"learning_rate": 0.0001600098089012244,
"loss": 0.9655,
"step": 1585
},
{
"epoch": 0.3668666359021689,
"grad_norm": 0.26328402757644653,
"learning_rate": 0.0001596871380295351,
"loss": 0.9608,
"step": 1590
},
{
"epoch": 0.3680203045685279,
"grad_norm": 0.27382150292396545,
"learning_rate": 0.00015936349890070602,
"loss": 0.9505,
"step": 1595
},
{
"epoch": 0.36917397323488693,
"grad_norm": 0.26438233256340027,
"learning_rate": 0.00015903889676487833,
"loss": 0.9227,
"step": 1600
},
{
"epoch": 0.37032764190124595,
"grad_norm": 0.2556706666946411,
"learning_rate": 0.00015871333688781522,
"loss": 0.9931,
"step": 1605
},
{
"epoch": 0.37148131056760497,
"grad_norm": 0.27087247371673584,
"learning_rate": 0.00015838682455081657,
"loss": 0.9914,
"step": 1610
},
{
"epoch": 0.372634979233964,
"grad_norm": 0.27084994316101074,
"learning_rate": 0.00015805936505063322,
"loss": 0.9641,
"step": 1615
},
{
"epoch": 0.373788647900323,
"grad_norm": 0.2637239396572113,
"learning_rate": 0.00015773096369938125,
"loss": 0.96,
"step": 1620
},
{
"epoch": 0.37494231656668203,
"grad_norm": 0.25956466794013977,
"learning_rate": 0.00015740162582445546,
"loss": 0.9617,
"step": 1625
},
{
"epoch": 0.37609598523304105,
"grad_norm": 0.2576027810573578,
"learning_rate": 0.0001570713567684432,
"loss": 0.9539,
"step": 1630
},
{
"epoch": 0.3772496538994001,
"grad_norm": 0.25418850779533386,
"learning_rate": 0.00015674016188903762,
"loss": 0.9145,
"step": 1635
},
{
"epoch": 0.3784033225657591,
"grad_norm": 0.2608237564563751,
"learning_rate": 0.00015640804655895084,
"loss": 0.9819,
"step": 1640
},
{
"epoch": 0.3795569912321181,
"grad_norm": 0.2650778889656067,
"learning_rate": 0.00015607501616582664,
"loss": 0.9296,
"step": 1645
},
{
"epoch": 0.38071065989847713,
"grad_norm": 0.2612312138080597,
"learning_rate": 0.00015574107611215319,
"loss": 0.9691,
"step": 1650
},
{
"epoch": 0.38186432856483615,
"grad_norm": 0.2598936855792999,
"learning_rate": 0.00015540623181517531,
"loss": 0.9305,
"step": 1655
},
{
"epoch": 0.3830179972311952,
"grad_norm": 0.26123687624931335,
"learning_rate": 0.00015507048870680668,
"loss": 0.9859,
"step": 1660
},
{
"epoch": 0.38417166589755425,
"grad_norm": 0.26335862278938293,
"learning_rate": 0.00015473385223354168,
"loss": 0.9645,
"step": 1665
},
{
"epoch": 0.38532533456391327,
"grad_norm": 0.26434382796287537,
"learning_rate": 0.00015439632785636706,
"loss": 0.9682,
"step": 1670
},
{
"epoch": 0.3864790032302723,
"grad_norm": 0.2613257169723511,
"learning_rate": 0.00015405792105067332,
"loss": 0.9602,
"step": 1675
},
{
"epoch": 0.3876326718966313,
"grad_norm": 0.2707611918449402,
"learning_rate": 0.00015371863730616586,
"loss": 0.9942,
"step": 1680
},
{
"epoch": 0.38878634056299033,
"grad_norm": 0.25917673110961914,
"learning_rate": 0.00015337848212677597,
"loss": 0.9372,
"step": 1685
},
{
"epoch": 0.38994000922934935,
"grad_norm": 0.27119165658950806,
"learning_rate": 0.00015303746103057162,
"loss": 0.9884,
"step": 1690
},
{
"epoch": 0.39109367789570837,
"grad_norm": 0.26019906997680664,
"learning_rate": 0.00015269557954966778,
"loss": 0.9684,
"step": 1695
},
{
"epoch": 0.3922473465620674,
"grad_norm": 0.2654115557670593,
"learning_rate": 0.00015235284323013675,
"loss": 0.955,
"step": 1700
},
{
"epoch": 0.3934010152284264,
"grad_norm": 0.26324862241744995,
"learning_rate": 0.0001520092576319183,
"loss": 0.9464,
"step": 1705
},
{
"epoch": 0.39455468389478543,
"grad_norm": 0.270831823348999,
"learning_rate": 0.00015166482832872923,
"loss": 0.9716,
"step": 1710
},
{
"epoch": 0.39570835256114445,
"grad_norm": 0.27920085191726685,
"learning_rate": 0.00015131956090797325,
"loss": 0.9674,
"step": 1715
},
{
"epoch": 0.39686202122750347,
"grad_norm": 0.27237191796302795,
"learning_rate": 0.00015097346097065007,
"loss": 1.0088,
"step": 1720
},
{
"epoch": 0.3980156898938625,
"grad_norm": 0.27462947368621826,
"learning_rate": 0.0001506265341312647,
"loss": 0.9321,
"step": 1725
},
{
"epoch": 0.3991693585602215,
"grad_norm": 0.26354286074638367,
"learning_rate": 0.00015027878601773633,
"loss": 0.9591,
"step": 1730
},
{
"epoch": 0.40032302722658053,
"grad_norm": 0.2659231126308441,
"learning_rate": 0.000149930222271307,
"loss": 0.9717,
"step": 1735
},
{
"epoch": 0.40147669589293955,
"grad_norm": 0.2720133662223816,
"learning_rate": 0.0001495808485464502,
"loss": 0.9727,
"step": 1740
},
{
"epoch": 0.40263036455929857,
"grad_norm": 0.25452399253845215,
"learning_rate": 0.00014923067051077893,
"loss": 0.9862,
"step": 1745
},
{
"epoch": 0.4037840332256576,
"grad_norm": 0.2635675072669983,
"learning_rate": 0.00014887969384495402,
"loss": 0.9508,
"step": 1750
},
{
"epoch": 0.4049377018920166,
"grad_norm": 0.2562003433704376,
"learning_rate": 0.0001485279242425917,
"loss": 0.9942,
"step": 1755
},
{
"epoch": 0.40609137055837563,
"grad_norm": 0.26377397775650024,
"learning_rate": 0.00014817536741017152,
"loss": 0.9425,
"step": 1760
},
{
"epoch": 0.40724503922473465,
"grad_norm": 0.2801288664340973,
"learning_rate": 0.0001478220290669436,
"loss": 0.9897,
"step": 1765
},
{
"epoch": 0.40839870789109367,
"grad_norm": 0.27296552062034607,
"learning_rate": 0.00014746791494483583,
"loss": 0.9771,
"step": 1770
},
{
"epoch": 0.4095523765574527,
"grad_norm": 0.27594876289367676,
"learning_rate": 0.000147113030788361,
"loss": 0.9926,
"step": 1775
},
{
"epoch": 0.4107060452238117,
"grad_norm": 0.2638562321662903,
"learning_rate": 0.00014675738235452352,
"loss": 0.9543,
"step": 1780
},
{
"epoch": 0.41185971389017073,
"grad_norm": 0.2608337998390198,
"learning_rate": 0.0001464009754127261,
"loss": 0.9612,
"step": 1785
},
{
"epoch": 0.41301338255652975,
"grad_norm": 0.2768416702747345,
"learning_rate": 0.00014604381574467615,
"loss": 0.9678,
"step": 1790
},
{
"epoch": 0.41416705122288877,
"grad_norm": 0.25971782207489014,
"learning_rate": 0.0001456859091442919,
"loss": 0.9335,
"step": 1795
},
{
"epoch": 0.4153207198892478,
"grad_norm": 0.25914978981018066,
"learning_rate": 0.00014532726141760848,
"loss": 1.0066,
"step": 1800
},
{
"epoch": 0.4164743885556068,
"grad_norm": 0.25939711928367615,
"learning_rate": 0.00014496787838268378,
"loss": 0.9687,
"step": 1805
},
{
"epoch": 0.41762805722196583,
"grad_norm": 0.2599487900733948,
"learning_rate": 0.00014460776586950393,
"loss": 0.9776,
"step": 1810
},
{
"epoch": 0.41878172588832485,
"grad_norm": 0.25991660356521606,
"learning_rate": 0.00014424692971988886,
"loss": 0.9543,
"step": 1815
},
{
"epoch": 0.41993539455468387,
"grad_norm": 0.2667793035507202,
"learning_rate": 0.0001438853757873975,
"loss": 0.944,
"step": 1820
},
{
"epoch": 0.4210890632210429,
"grad_norm": 0.2818945348262787,
"learning_rate": 0.00014352310993723277,
"loss": 0.9587,
"step": 1825
},
{
"epoch": 0.4222427318874019,
"grad_norm": 0.271699994802475,
"learning_rate": 0.00014316013804614643,
"loss": 0.9647,
"step": 1830
},
{
"epoch": 0.423396400553761,
"grad_norm": 0.2668125331401825,
"learning_rate": 0.00014279646600234386,
"loss": 0.9732,
"step": 1835
},
{
"epoch": 0.42455006922012,
"grad_norm": 0.25676923990249634,
"learning_rate": 0.00014243209970538846,
"loss": 0.9776,
"step": 1840
},
{
"epoch": 0.42570373788647903,
"grad_norm": 0.2696235775947571,
"learning_rate": 0.00014206704506610583,
"loss": 0.9644,
"step": 1845
},
{
"epoch": 0.42685740655283805,
"grad_norm": 0.25627243518829346,
"learning_rate": 0.00014170130800648814,
"loss": 0.9586,
"step": 1850
},
{
"epoch": 0.42801107521919707,
"grad_norm": 0.2514009475708008,
"learning_rate": 0.00014133489445959787,
"loss": 0.9311,
"step": 1855
},
{
"epoch": 0.4291647438855561,
"grad_norm": 0.24791298806667328,
"learning_rate": 0.00014096781036947157,
"loss": 0.9819,
"step": 1860
},
{
"epoch": 0.4303184125519151,
"grad_norm": 0.2555059492588043,
"learning_rate": 0.00014060006169102363,
"loss": 0.9501,
"step": 1865
},
{
"epoch": 0.43147208121827413,
"grad_norm": 0.25847703218460083,
"learning_rate": 0.0001402316543899493,
"loss": 0.9871,
"step": 1870
},
{
"epoch": 0.43262574988463315,
"grad_norm": 0.2684321701526642,
"learning_rate": 0.0001398625944426284,
"loss": 0.9871,
"step": 1875
},
{
"epoch": 0.43377941855099217,
"grad_norm": 0.2601282000541687,
"learning_rate": 0.0001394928878360279,
"loss": 1.0022,
"step": 1880
},
{
"epoch": 0.4349330872173512,
"grad_norm": 0.26379039883613586,
"learning_rate": 0.0001391225405676051,
"loss": 0.9693,
"step": 1885
},
{
"epoch": 0.4360867558837102,
"grad_norm": 0.2606624662876129,
"learning_rate": 0.0001387515586452103,
"loss": 0.9585,
"step": 1890
},
{
"epoch": 0.43724042455006923,
"grad_norm": 0.269196480512619,
"learning_rate": 0.0001383799480869892,
"loss": 0.9539,
"step": 1895
},
{
"epoch": 0.43839409321642825,
"grad_norm": 0.26253777742385864,
"learning_rate": 0.00013800771492128536,
"loss": 0.9294,
"step": 1900
},
{
"epoch": 0.43954776188278727,
"grad_norm": 0.2613804340362549,
"learning_rate": 0.00013763486518654253,
"loss": 0.9316,
"step": 1905
},
{
"epoch": 0.4407014305491463,
"grad_norm": 0.25759026408195496,
"learning_rate": 0.0001372614049312064,
"loss": 0.9558,
"step": 1910
},
{
"epoch": 0.4418550992155053,
"grad_norm": 0.25641101598739624,
"learning_rate": 0.00013688734021362675,
"loss": 0.9405,
"step": 1915
},
{
"epoch": 0.44300876788186433,
"grad_norm": 0.26334041357040405,
"learning_rate": 0.00013651267710195907,
"loss": 0.9624,
"step": 1920
},
{
"epoch": 0.44416243654822335,
"grad_norm": 0.2584952712059021,
"learning_rate": 0.00013613742167406612,
"loss": 0.9953,
"step": 1925
},
{
"epoch": 0.44531610521458237,
"grad_norm": 0.2532113790512085,
"learning_rate": 0.00013576158001741932,
"loss": 0.942,
"step": 1930
},
{
"epoch": 0.4464697738809414,
"grad_norm": 0.25842392444610596,
"learning_rate": 0.00013538515822900002,
"loss": 0.941,
"step": 1935
},
{
"epoch": 0.4476234425473004,
"grad_norm": 0.27943527698516846,
"learning_rate": 0.00013500816241520058,
"loss": 0.9809,
"step": 1940
},
{
"epoch": 0.44877711121365943,
"grad_norm": 0.2741059362888336,
"learning_rate": 0.00013463059869172535,
"loss": 0.9559,
"step": 1945
},
{
"epoch": 0.44993077988001845,
"grad_norm": 0.27189043164253235,
"learning_rate": 0.00013425247318349137,
"loss": 0.9496,
"step": 1950
},
{
"epoch": 0.45108444854637747,
"grad_norm": 0.2595981955528259,
"learning_rate": 0.00013387379202452917,
"loss": 0.9404,
"step": 1955
},
{
"epoch": 0.4522381172127365,
"grad_norm": 0.2670726180076599,
"learning_rate": 0.00013349456135788298,
"loss": 0.9933,
"step": 1960
},
{
"epoch": 0.4533917858790955,
"grad_norm": 0.26899731159210205,
"learning_rate": 0.0001331147873355115,
"loss": 0.9498,
"step": 1965
},
{
"epoch": 0.45454545454545453,
"grad_norm": 0.25968730449676514,
"learning_rate": 0.00013273447611818767,
"loss": 0.9623,
"step": 1970
},
{
"epoch": 0.45569912321181355,
"grad_norm": 0.2737217843532562,
"learning_rate": 0.000132353633875399,
"loss": 0.9661,
"step": 1975
},
{
"epoch": 0.45685279187817257,
"grad_norm": 0.2898109555244446,
"learning_rate": 0.00013197226678524738,
"loss": 0.9716,
"step": 1980
},
{
"epoch": 0.4580064605445316,
"grad_norm": 0.27403295040130615,
"learning_rate": 0.00013159038103434888,
"loss": 0.9601,
"step": 1985
},
{
"epoch": 0.4591601292108906,
"grad_norm": 0.2780100405216217,
"learning_rate": 0.00013120798281773347,
"loss": 0.9191,
"step": 1990
},
{
"epoch": 0.46031379787724963,
"grad_norm": 0.25929951667785645,
"learning_rate": 0.00013082507833874426,
"loss": 0.9445,
"step": 1995
},
{
"epoch": 0.46146746654360865,
"grad_norm": 0.2623712420463562,
"learning_rate": 0.00013044167380893727,
"loss": 0.9961,
"step": 2000
},
{
"epoch": 0.46262113520996767,
"grad_norm": 0.27212023735046387,
"learning_rate": 0.00013005777544798026,
"loss": 0.9497,
"step": 2005
},
{
"epoch": 0.46377480387632675,
"grad_norm": 0.2612648606300354,
"learning_rate": 0.00012967338948355217,
"loss": 0.9477,
"step": 2010
},
{
"epoch": 0.46492847254268577,
"grad_norm": 0.27296924591064453,
"learning_rate": 0.0001292885221512419,
"loss": 0.9788,
"step": 2015
},
{
"epoch": 0.4660821412090448,
"grad_norm": 0.24790716171264648,
"learning_rate": 0.00012890317969444716,
"loss": 0.942,
"step": 2020
},
{
"epoch": 0.4672358098754038,
"grad_norm": 0.27568113803863525,
"learning_rate": 0.00012851736836427325,
"loss": 0.9746,
"step": 2025
},
{
"epoch": 0.4683894785417628,
"grad_norm": 0.2629447281360626,
"learning_rate": 0.00012813109441943166,
"loss": 0.9481,
"step": 2030
},
{
"epoch": 0.46954314720812185,
"grad_norm": 0.264850914478302,
"learning_rate": 0.00012774436412613845,
"loss": 0.9723,
"step": 2035
},
{
"epoch": 0.47069681587448087,
"grad_norm": 0.27910053730010986,
"learning_rate": 0.0001273571837580127,
"loss": 0.9671,
"step": 2040
},
{
"epoch": 0.4718504845408399,
"grad_norm": 0.26526764035224915,
"learning_rate": 0.0001269695595959747,
"loss": 0.9891,
"step": 2045
},
{
"epoch": 0.4730041532071989,
"grad_norm": 0.27195367217063904,
"learning_rate": 0.00012658149792814404,
"loss": 0.9624,
"step": 2050
},
{
"epoch": 0.4741578218735579,
"grad_norm": 0.2728367745876312,
"learning_rate": 0.00012619300504973762,
"loss": 0.9487,
"step": 2055
},
{
"epoch": 0.47531149053991695,
"grad_norm": 0.2927708625793457,
"learning_rate": 0.0001258040872629676,
"loss": 0.9206,
"step": 2060
},
{
"epoch": 0.47646515920627597,
"grad_norm": 0.2730552852153778,
"learning_rate": 0.00012541475087693896,
"loss": 0.9995,
"step": 2065
},
{
"epoch": 0.477618827872635,
"grad_norm": 0.26003938913345337,
"learning_rate": 0.00012502500220754737,
"loss": 0.9627,
"step": 2070
},
{
"epoch": 0.478772496538994,
"grad_norm": 0.26204952597618103,
"learning_rate": 0.00012463484757737662,
"loss": 0.959,
"step": 2075
},
{
"epoch": 0.47992616520535303,
"grad_norm": 0.25478288531303406,
"learning_rate": 0.0001242442933155961,
"loss": 0.9581,
"step": 2080
},
{
"epoch": 0.48107983387171205,
"grad_norm": 0.2646753489971161,
"learning_rate": 0.0001238533457578581,
"loss": 0.9405,
"step": 2085
},
{
"epoch": 0.48223350253807107,
"grad_norm": 0.2728932201862335,
"learning_rate": 0.00012346201124619502,
"loss": 0.9375,
"step": 2090
},
{
"epoch": 0.4833871712044301,
"grad_norm": 0.26134753227233887,
"learning_rate": 0.00012307029612891655,
"loss": 0.9525,
"step": 2095
},
{
"epoch": 0.4845408398707891,
"grad_norm": 0.2668938636779785,
"learning_rate": 0.00012267820676050656,
"loss": 0.9239,
"step": 2100
},
{
"epoch": 0.48569450853714813,
"grad_norm": 0.25549912452697754,
"learning_rate": 0.00012228574950152017,
"loss": 0.9374,
"step": 2105
},
{
"epoch": 0.48684817720350715,
"grad_norm": 0.5779925584793091,
"learning_rate": 0.00012189293071848051,
"loss": 0.9765,
"step": 2110
},
{
"epoch": 0.48800184586986617,
"grad_norm": 0.25671547651290894,
"learning_rate": 0.00012149975678377541,
"loss": 1.0103,
"step": 2115
},
{
"epoch": 0.4891555145362252,
"grad_norm": 0.2744820713996887,
"learning_rate": 0.00012110623407555397,
"loss": 0.9772,
"step": 2120
},
{
"epoch": 0.4903091832025842,
"grad_norm": 0.2585495114326477,
"learning_rate": 0.00012071236897762325,
"loss": 0.9565,
"step": 2125
},
{
"epoch": 0.49146285186894323,
"grad_norm": 0.2611427307128906,
"learning_rate": 0.00012031816787934464,
"loss": 0.9505,
"step": 2130
},
{
"epoch": 0.49261652053530225,
"grad_norm": 0.2629440426826477,
"learning_rate": 0.00011992363717553015,
"loss": 0.9399,
"step": 2135
},
{
"epoch": 0.49377018920166127,
"grad_norm": 0.25579625368118286,
"learning_rate": 0.00011952878326633872,
"loss": 0.8824,
"step": 2140
},
{
"epoch": 0.4949238578680203,
"grad_norm": 0.26876822113990784,
"learning_rate": 0.00011913361255717241,
"loss": 0.949,
"step": 2145
},
{
"epoch": 0.4960775265343793,
"grad_norm": 0.2676302194595337,
"learning_rate": 0.00011873813145857249,
"loss": 0.9855,
"step": 2150
},
{
"epoch": 0.49723119520073833,
"grad_norm": 0.26593196392059326,
"learning_rate": 0.00011834234638611539,
"loss": 0.9051,
"step": 2155
},
{
"epoch": 0.49838486386709735,
"grad_norm": 0.2615763247013092,
"learning_rate": 0.00011794626376030866,
"loss": 0.9362,
"step": 2160
},
{
"epoch": 0.49953853253345637,
"grad_norm": 0.260310560464859,
"learning_rate": 0.00011754989000648693,
"loss": 0.9614,
"step": 2165
},
{
"epoch": 0.5006922011998154,
"grad_norm": 0.26325076818466187,
"learning_rate": 0.00011715323155470745,
"loss": 0.9321,
"step": 2170
},
{
"epoch": 0.5018458698661744,
"grad_norm": 0.2704361081123352,
"learning_rate": 0.00011675629483964596,
"loss": 0.933,
"step": 2175
},
{
"epoch": 0.5029995385325334,
"grad_norm": 0.2598739266395569,
"learning_rate": 0.0001163590863004922,
"loss": 0.9706,
"step": 2180
},
{
"epoch": 0.5041532071988925,
"grad_norm": 0.2715523838996887,
"learning_rate": 0.0001159616123808455,
"loss": 0.9422,
"step": 2185
},
{
"epoch": 0.5053068758652515,
"grad_norm": 0.2593281865119934,
"learning_rate": 0.00011556387952861036,
"loss": 1.0014,
"step": 2190
},
{
"epoch": 0.5064605445316105,
"grad_norm": 0.26070570945739746,
"learning_rate": 0.0001151658941958916,
"loss": 0.9399,
"step": 2195
},
{
"epoch": 0.5076142131979695,
"grad_norm": 0.2727888524532318,
"learning_rate": 0.00011476766283888986,
"loss": 0.9557,
"step": 2200
},
{
"epoch": 0.5087678818643285,
"grad_norm": 0.2623576819896698,
"learning_rate": 0.00011436919191779687,
"loss": 0.976,
"step": 2205
},
{
"epoch": 0.5099215505306876,
"grad_norm": 0.26719269156455994,
"learning_rate": 0.0001139704878966906,
"loss": 0.9823,
"step": 2210
},
{
"epoch": 0.5110752191970466,
"grad_norm": 0.2678567171096802,
"learning_rate": 0.00011357155724343045,
"loss": 0.9361,
"step": 2215
},
{
"epoch": 0.5122288878634056,
"grad_norm": 0.2595786154270172,
"learning_rate": 0.00011317240642955225,
"loss": 0.9625,
"step": 2220
},
{
"epoch": 0.5133825565297646,
"grad_norm": 0.2738422751426697,
"learning_rate": 0.00011277304193016332,
"loss": 0.985,
"step": 2225
},
{
"epoch": 0.5145362251961236,
"grad_norm": 0.2674945890903473,
"learning_rate": 0.00011237347022383746,
"loss": 0.9513,
"step": 2230
},
{
"epoch": 0.5156898938624827,
"grad_norm": 0.2577970325946808,
"learning_rate": 0.00011197369779250979,
"loss": 0.9865,
"step": 2235
},
{
"epoch": 0.5168435625288417,
"grad_norm": 0.27681398391723633,
"learning_rate": 0.00011157373112137171,
"loss": 0.9499,
"step": 2240
},
{
"epoch": 0.5179972311952007,
"grad_norm": 0.27841857075691223,
"learning_rate": 0.0001111735766987655,
"loss": 0.9625,
"step": 2245
},
{
"epoch": 0.5191508998615597,
"grad_norm": 0.26672080159187317,
"learning_rate": 0.00011077324101607929,
"loss": 0.9517,
"step": 2250
},
{
"epoch": 0.5203045685279187,
"grad_norm": 0.2566830813884735,
"learning_rate": 0.00011037273056764157,
"loss": 0.9508,
"step": 2255
},
{
"epoch": 0.5214582371942778,
"grad_norm": 0.2679496705532074,
"learning_rate": 0.00010997205185061599,
"loss": 0.9332,
"step": 2260
},
{
"epoch": 0.5226119058606368,
"grad_norm": 0.2658250629901886,
"learning_rate": 0.00010957121136489581,
"loss": 0.95,
"step": 2265
},
{
"epoch": 0.5237655745269958,
"grad_norm": 0.2708585560321808,
"learning_rate": 0.00010917021561299863,
"loss": 0.9577,
"step": 2270
},
{
"epoch": 0.5249192431933549,
"grad_norm": 0.25992119312286377,
"learning_rate": 0.0001087690710999607,
"loss": 0.969,
"step": 2275
},
{
"epoch": 0.5260729118597139,
"grad_norm": 0.2593708336353302,
"learning_rate": 0.00010836778433323158,
"loss": 0.9516,
"step": 2280
},
{
"epoch": 0.527226580526073,
"grad_norm": 0.267589807510376,
"learning_rate": 0.00010796636182256845,
"loss": 0.9257,
"step": 2285
},
{
"epoch": 0.528380249192432,
"grad_norm": 0.26011785864830017,
"learning_rate": 0.00010756481007993063,
"loss": 0.994,
"step": 2290
},
{
"epoch": 0.529533917858791,
"grad_norm": 0.27399560809135437,
"learning_rate": 0.0001071631356193738,
"loss": 0.9628,
"step": 2295
},
{
"epoch": 0.53068758652515,
"grad_norm": 0.25914278626441956,
"learning_rate": 0.00010676134495694439,
"loss": 0.9675,
"step": 2300
},
{
"epoch": 0.531841255191509,
"grad_norm": 0.26276910305023193,
"learning_rate": 0.00010635944461057395,
"loss": 0.9666,
"step": 2305
},
{
"epoch": 0.5329949238578681,
"grad_norm": 0.26222512125968933,
"learning_rate": 0.00010595744109997325,
"loss": 0.9434,
"step": 2310
},
{
"epoch": 0.5341485925242271,
"grad_norm": 0.269531786441803,
"learning_rate": 0.00010555534094652675,
"loss": 0.9613,
"step": 2315
},
{
"epoch": 0.5353022611905861,
"grad_norm": 0.2551771104335785,
"learning_rate": 0.00010515315067318652,
"loss": 0.9855,
"step": 2320
},
{
"epoch": 0.5364559298569451,
"grad_norm": 0.25686752796173096,
"learning_rate": 0.00010475087680436666,
"loss": 0.9563,
"step": 2325
},
{
"epoch": 0.5376095985233041,
"grad_norm": 0.2565601170063019,
"learning_rate": 0.00010434852586583736,
"loss": 0.966,
"step": 2330
},
{
"epoch": 0.5387632671896632,
"grad_norm": 0.2918190360069275,
"learning_rate": 0.000103946104384619,
"loss": 0.9314,
"step": 2335
},
{
"epoch": 0.5399169358560222,
"grad_norm": 0.2644803822040558,
"learning_rate": 0.00010354361888887642,
"loss": 0.9581,
"step": 2340
},
{
"epoch": 0.5410706045223812,
"grad_norm": 0.2727797031402588,
"learning_rate": 0.00010314107590781284,
"loss": 0.9633,
"step": 2345
},
{
"epoch": 0.5422242731887402,
"grad_norm": 0.273027241230011,
"learning_rate": 0.00010273848197156401,
"loss": 0.9487,
"step": 2350
},
{
"epoch": 0.5433779418550992,
"grad_norm": 0.2540397644042969,
"learning_rate": 0.00010233584361109235,
"loss": 0.9735,
"step": 2355
},
{
"epoch": 0.5445316105214583,
"grad_norm": 0.2538515031337738,
"learning_rate": 0.00010193316735808085,
"loss": 0.9788,
"step": 2360
},
{
"epoch": 0.5456852791878173,
"grad_norm": 0.27701374888420105,
"learning_rate": 0.00010153045974482732,
"loss": 0.9557,
"step": 2365
},
{
"epoch": 0.5468389478541763,
"grad_norm": 0.25808069109916687,
"learning_rate": 0.00010112772730413815,
"loss": 0.9243,
"step": 2370
},
{
"epoch": 0.5479926165205353,
"grad_norm": 0.256411075592041,
"learning_rate": 0.00010072497656922266,
"loss": 0.9752,
"step": 2375
},
{
"epoch": 0.5491462851868943,
"grad_norm": 0.2654629647731781,
"learning_rate": 0.00010032221407358681,
"loss": 0.9949,
"step": 2380
},
{
"epoch": 0.5502999538532534,
"grad_norm": 0.2696619927883148,
"learning_rate": 9.99194463509274e-05,
"loss": 0.971,
"step": 2385
},
{
"epoch": 0.5514536225196124,
"grad_norm": 0.2578607499599457,
"learning_rate": 9.9516679935026e-05,
"loss": 0.9533,
"step": 2390
},
{
"epoch": 0.5526072911859714,
"grad_norm": 0.26021629571914673,
"learning_rate": 9.911392135964298e-05,
"loss": 0.9677,
"step": 2395
},
{
"epoch": 0.5537609598523304,
"grad_norm": 0.27699634432792664,
"learning_rate": 9.871117715841151e-05,
"loss": 0.9434,
"step": 2400
},
{
"epoch": 0.5549146285186894,
"grad_norm": 0.25480917096138,
"learning_rate": 9.830845386473168e-05,
"loss": 0.9836,
"step": 2405
},
{
"epoch": 0.5560682971850485,
"grad_norm": 0.2536785304546356,
"learning_rate": 9.790575801166432e-05,
"loss": 0.9518,
"step": 2410
},
{
"epoch": 0.5572219658514075,
"grad_norm": 0.2608122229576111,
"learning_rate": 9.750309613182505e-05,
"loss": 0.9649,
"step": 2415
},
{
"epoch": 0.5583756345177665,
"grad_norm": 0.25869712233543396,
"learning_rate": 9.710047475727855e-05,
"loss": 0.9668,
"step": 2420
},
{
"epoch": 0.5595293031841255,
"grad_norm": 0.25615230202674866,
"learning_rate": 9.669790041943225e-05,
"loss": 0.9062,
"step": 2425
},
{
"epoch": 0.5606829718504845,
"grad_norm": 0.26828473806381226,
"learning_rate": 9.629537964893063e-05,
"loss": 0.9787,
"step": 2430
},
{
"epoch": 0.5618366405168436,
"grad_norm": 0.25551798939704895,
"learning_rate": 9.589291897554912e-05,
"loss": 0.9597,
"step": 2435
},
{
"epoch": 0.5629903091832026,
"grad_norm": 0.26706936955451965,
"learning_rate": 9.549052492808834e-05,
"loss": 0.9728,
"step": 2440
},
{
"epoch": 0.5641439778495616,
"grad_norm": 0.25929680466651917,
"learning_rate": 9.508820403426799e-05,
"loss": 0.9705,
"step": 2445
},
{
"epoch": 0.5652976465159206,
"grad_norm": 2.2861907482147217,
"learning_rate": 9.468596282062114e-05,
"loss": 0.9766,
"step": 2450
},
{
"epoch": 0.5664513151822796,
"grad_norm": 0.2641540467739105,
"learning_rate": 9.428380781238821e-05,
"loss": 1.0021,
"step": 2455
},
{
"epoch": 0.5676049838486387,
"grad_norm": 0.25216275453567505,
"learning_rate": 9.38817455334112e-05,
"loss": 0.9563,
"step": 2460
},
{
"epoch": 0.5687586525149977,
"grad_norm": 0.262725293636322,
"learning_rate": 9.347978250602785e-05,
"loss": 0.9765,
"step": 2465
},
{
"epoch": 0.5699123211813567,
"grad_norm": 0.2671465277671814,
"learning_rate": 9.307792525096581e-05,
"loss": 0.9308,
"step": 2470
},
{
"epoch": 0.5710659898477157,
"grad_norm": 0.26376160979270935,
"learning_rate": 9.267618028723686e-05,
"loss": 0.9807,
"step": 2475
},
{
"epoch": 0.5722196585140747,
"grad_norm": 0.24652628600597382,
"learning_rate": 9.227455413203115e-05,
"loss": 0.9261,
"step": 2480
},
{
"epoch": 0.5733733271804338,
"grad_norm": 0.2701497972011566,
"learning_rate": 9.187305330061156e-05,
"loss": 0.9542,
"step": 2485
},
{
"epoch": 0.5745269958467928,
"grad_norm": 0.28991737961769104,
"learning_rate": 9.147168430620787e-05,
"loss": 0.9431,
"step": 2490
},
{
"epoch": 0.5756806645131518,
"grad_norm": 0.2657528817653656,
"learning_rate": 9.107045365991123e-05,
"loss": 0.9601,
"step": 2495
},
{
"epoch": 0.5768343331795108,
"grad_norm": 0.25293365120887756,
"learning_rate": 9.066936787056842e-05,
"loss": 0.9528,
"step": 2500
},
{
"epoch": 0.5779880018458698,
"grad_norm": 0.26489248871803284,
"learning_rate": 9.026843344467635e-05,
"loss": 0.9509,
"step": 2505
},
{
"epoch": 0.5791416705122289,
"grad_norm": 0.27219951152801514,
"learning_rate": 8.986765688627652e-05,
"loss": 0.9485,
"step": 2510
},
{
"epoch": 0.5802953391785879,
"grad_norm": 0.2632032334804535,
"learning_rate": 8.946704469684939e-05,
"loss": 0.9635,
"step": 2515
},
{
"epoch": 0.5814490078449469,
"grad_norm": 0.25340795516967773,
"learning_rate": 8.906660337520903e-05,
"loss": 0.9315,
"step": 2520
},
{
"epoch": 0.5826026765113059,
"grad_norm": 0.26531898975372314,
"learning_rate": 8.86663394173977e-05,
"loss": 0.974,
"step": 2525
},
{
"epoch": 0.583756345177665,
"grad_norm": 0.26169174909591675,
"learning_rate": 8.826625931658039e-05,
"loss": 0.9416,
"step": 2530
},
{
"epoch": 0.584910013844024,
"grad_norm": 0.27198314666748047,
"learning_rate": 8.786636956293948e-05,
"loss": 0.9827,
"step": 2535
},
{
"epoch": 0.586063682510383,
"grad_norm": 0.2525466978549957,
"learning_rate": 8.746667664356956e-05,
"loss": 0.9506,
"step": 2540
},
{
"epoch": 0.587217351176742,
"grad_norm": 0.2605235278606415,
"learning_rate": 8.706718704237215e-05,
"loss": 0.9747,
"step": 2545
},
{
"epoch": 0.588371019843101,
"grad_norm": 0.263271301984787,
"learning_rate": 8.666790723995042e-05,
"loss": 0.9495,
"step": 2550
},
{
"epoch": 0.58952468850946,
"grad_norm": 0.2623215317726135,
"learning_rate": 8.626884371350421e-05,
"loss": 0.9407,
"step": 2555
},
{
"epoch": 0.5906783571758191,
"grad_norm": 0.2711637616157532,
"learning_rate": 8.587000293672481e-05,
"loss": 0.9332,
"step": 2560
},
{
"epoch": 0.5918320258421781,
"grad_norm": 0.27216318249702454,
"learning_rate": 8.547139137969015e-05,
"loss": 0.9789,
"step": 2565
},
{
"epoch": 0.5929856945085371,
"grad_norm": 0.2510489821434021,
"learning_rate": 8.50730155087596e-05,
"loss": 0.9522,
"step": 2570
},
{
"epoch": 0.5941393631748961,
"grad_norm": 0.27893054485321045,
"learning_rate": 8.46748817864692e-05,
"loss": 0.9849,
"step": 2575
},
{
"epoch": 0.5952930318412551,
"grad_norm": 0.2612435817718506,
"learning_rate": 8.427699667142682e-05,
"loss": 0.9512,
"step": 2580
},
{
"epoch": 0.5964467005076142,
"grad_norm": 0.26378709077835083,
"learning_rate": 8.387936661820734e-05,
"loss": 0.8909,
"step": 2585
},
{
"epoch": 0.5976003691739732,
"grad_norm": 0.2865258455276489,
"learning_rate": 8.348199807724806e-05,
"loss": 0.964,
"step": 2590
},
{
"epoch": 0.5987540378403322,
"grad_norm": 0.26403477787971497,
"learning_rate": 8.308489749474388e-05,
"loss": 0.9232,
"step": 2595
},
{
"epoch": 0.5999077065066912,
"grad_norm": 0.2571257948875427,
"learning_rate": 8.268807131254287e-05,
"loss": 0.935,
"step": 2600
},
{
"epoch": 0.6010613751730502,
"grad_norm": 0.2499091774225235,
"learning_rate": 8.229152596804168e-05,
"loss": 0.9262,
"step": 2605
},
{
"epoch": 0.6022150438394093,
"grad_norm": 0.2728745639324188,
"learning_rate": 8.189526789408123e-05,
"loss": 0.984,
"step": 2610
},
{
"epoch": 0.6033687125057683,
"grad_norm": 0.27502092719078064,
"learning_rate": 8.149930351884221e-05,
"loss": 0.9326,
"step": 2615
},
{
"epoch": 0.6045223811721273,
"grad_norm": 0.26331180334091187,
"learning_rate": 8.110363926574087e-05,
"loss": 0.9026,
"step": 2620
},
{
"epoch": 0.6056760498384864,
"grad_norm": 0.25487685203552246,
"learning_rate": 8.070828155332486e-05,
"loss": 0.9307,
"step": 2625
},
{
"epoch": 0.6068297185048455,
"grad_norm": 0.2543911933898926,
"learning_rate": 8.0313236795169e-05,
"loss": 0.9647,
"step": 2630
},
{
"epoch": 0.6079833871712045,
"grad_norm": 0.2619852125644684,
"learning_rate": 7.991851139977138e-05,
"loss": 0.9564,
"step": 2635
},
{
"epoch": 0.6091370558375635,
"grad_norm": 0.2640511989593506,
"learning_rate": 7.952411177044923e-05,
"loss": 0.9414,
"step": 2640
},
{
"epoch": 0.6102907245039225,
"grad_norm": 0.27362698316574097,
"learning_rate": 7.913004430523526e-05,
"loss": 0.9517,
"step": 2645
},
{
"epoch": 0.6114443931702815,
"grad_norm": 0.3130943477153778,
"learning_rate": 7.873631539677364e-05,
"loss": 0.9368,
"step": 2650
},
{
"epoch": 0.6125980618366406,
"grad_norm": 0.2534888684749603,
"learning_rate": 7.834293143221642e-05,
"loss": 0.9621,
"step": 2655
},
{
"epoch": 0.6137517305029996,
"grad_norm": 0.2623361349105835,
"learning_rate": 7.794989879311991e-05,
"loss": 0.9424,
"step": 2660
},
{
"epoch": 0.6149053991693586,
"grad_norm": 0.2610420882701874,
"learning_rate": 7.755722385534111e-05,
"loss": 0.9385,
"step": 2665
},
{
"epoch": 0.6160590678357176,
"grad_norm": 0.27034202218055725,
"learning_rate": 7.716491298893442e-05,
"loss": 0.9798,
"step": 2670
},
{
"epoch": 0.6172127365020766,
"grad_norm": 0.2572305202484131,
"learning_rate": 7.677297255804811e-05,
"loss": 0.9163,
"step": 2675
},
{
"epoch": 0.6183664051684357,
"grad_norm": 0.2526620626449585,
"learning_rate": 7.638140892082117e-05,
"loss": 0.9339,
"step": 2680
},
{
"epoch": 0.6195200738347947,
"grad_norm": 0.26071593165397644,
"learning_rate": 7.599022842928017e-05,
"loss": 0.9419,
"step": 2685
},
{
"epoch": 0.6206737425011537,
"grad_norm": 0.25418540835380554,
"learning_rate": 7.559943742923626e-05,
"loss": 0.9372,
"step": 2690
},
{
"epoch": 0.6218274111675127,
"grad_norm": 0.2598932385444641,
"learning_rate": 7.520904226018213e-05,
"loss": 0.9561,
"step": 2695
},
{
"epoch": 0.6229810798338717,
"grad_norm": 0.27341189980506897,
"learning_rate": 7.48190492551892e-05,
"loss": 0.9753,
"step": 2700
},
{
"epoch": 0.6241347485002308,
"grad_norm": 0.2658282518386841,
"learning_rate": 7.442946474080499e-05,
"loss": 0.9653,
"step": 2705
},
{
"epoch": 0.6252884171665898,
"grad_norm": 0.2684546411037445,
"learning_rate": 7.404029503695028e-05,
"loss": 0.9748,
"step": 2710
},
{
"epoch": 0.6264420858329488,
"grad_norm": 0.28556227684020996,
"learning_rate": 7.365154645681681e-05,
"loss": 0.9599,
"step": 2715
},
{
"epoch": 0.6275957544993078,
"grad_norm": 0.2757733166217804,
"learning_rate": 7.32632253067647e-05,
"loss": 0.9678,
"step": 2720
},
{
"epoch": 0.6287494231656668,
"grad_norm": 0.25145432353019714,
"learning_rate": 7.287533788622025e-05,
"loss": 0.9201,
"step": 2725
},
{
"epoch": 0.6299030918320259,
"grad_norm": 0.2673650085926056,
"learning_rate": 7.248789048757368e-05,
"loss": 0.9538,
"step": 2730
},
{
"epoch": 0.6310567604983849,
"grad_norm": 0.2711561918258667,
"learning_rate": 7.210088939607708e-05,
"loss": 0.9657,
"step": 2735
},
{
"epoch": 0.6322104291647439,
"grad_norm": 0.2643033266067505,
"learning_rate": 7.171434088974251e-05,
"loss": 0.9253,
"step": 2740
},
{
"epoch": 0.6333640978311029,
"grad_norm": 0.2604006826877594,
"learning_rate": 7.132825123924006e-05,
"loss": 0.9299,
"step": 2745
},
{
"epoch": 0.6345177664974619,
"grad_norm": 0.26541584730148315,
"learning_rate": 7.094262670779612e-05,
"loss": 0.9345,
"step": 2750
},
{
"epoch": 0.635671435163821,
"grad_norm": 0.2622869610786438,
"learning_rate": 7.055747355109186e-05,
"loss": 0.9559,
"step": 2755
},
{
"epoch": 0.63682510383018,
"grad_norm": 0.27132776379585266,
"learning_rate": 7.017279801716177e-05,
"loss": 0.9611,
"step": 2760
},
{
"epoch": 0.637978772496539,
"grad_norm": 0.25656870007514954,
"learning_rate": 6.978860634629214e-05,
"loss": 0.9387,
"step": 2765
},
{
"epoch": 0.639132441162898,
"grad_norm": 0.26165375113487244,
"learning_rate": 6.940490477092004e-05,
"loss": 0.9169,
"step": 2770
},
{
"epoch": 0.640286109829257,
"grad_norm": 0.24951021373271942,
"learning_rate": 6.902169951553202e-05,
"loss": 0.9349,
"step": 2775
},
{
"epoch": 0.6414397784956161,
"grad_norm": 0.2729644775390625,
"learning_rate": 6.863899679656328e-05,
"loss": 0.9506,
"step": 2780
},
{
"epoch": 0.6425934471619751,
"grad_norm": 0.2613593339920044,
"learning_rate": 6.82568028222967e-05,
"loss": 0.9612,
"step": 2785
},
{
"epoch": 0.6437471158283341,
"grad_norm": 0.254482626914978,
"learning_rate": 6.787512379276229e-05,
"loss": 0.9336,
"step": 2790
},
{
"epoch": 0.6449007844946931,
"grad_norm": 0.2548394501209259,
"learning_rate": 6.749396589963648e-05,
"loss": 0.9495,
"step": 2795
},
{
"epoch": 0.6460544531610521,
"grad_norm": 0.25152695178985596,
"learning_rate": 6.711333532614168e-05,
"loss": 0.9476,
"step": 2800
},
{
"epoch": 0.6472081218274112,
"grad_norm": 0.2617679536342621,
"learning_rate": 6.673323824694606e-05,
"loss": 0.9498,
"step": 2805
},
{
"epoch": 0.6483617904937702,
"grad_norm": 0.26209336519241333,
"learning_rate": 6.63536808280633e-05,
"loss": 0.9377,
"step": 2810
},
{
"epoch": 0.6495154591601292,
"grad_norm": 0.2582489848136902,
"learning_rate": 6.597466922675266e-05,
"loss": 0.9341,
"step": 2815
},
{
"epoch": 0.6506691278264882,
"grad_norm": 0.25654134154319763,
"learning_rate": 6.559620959141897e-05,
"loss": 0.922,
"step": 2820
},
{
"epoch": 0.6518227964928472,
"grad_norm": 0.25131335854530334,
"learning_rate": 6.521830806151297e-05,
"loss": 0.9427,
"step": 2825
},
{
"epoch": 0.6529764651592063,
"grad_norm": 0.26254597306251526,
"learning_rate": 6.48409707674317e-05,
"loss": 0.9359,
"step": 2830
},
{
"epoch": 0.6541301338255653,
"grad_norm": 0.2668085992336273,
"learning_rate": 6.446420383041903e-05,
"loss": 0.9273,
"step": 2835
},
{
"epoch": 0.6552838024919243,
"grad_norm": 0.24858598411083221,
"learning_rate": 6.408801336246645e-05,
"loss": 0.9611,
"step": 2840
},
{
"epoch": 0.6564374711582833,
"grad_norm": 0.2663392126560211,
"learning_rate": 6.371240546621378e-05,
"loss": 0.9497,
"step": 2845
},
{
"epoch": 0.6575911398246423,
"grad_norm": 0.26316505670547485,
"learning_rate": 6.333738623485025e-05,
"loss": 0.9674,
"step": 2850
},
{
"epoch": 0.6587448084910014,
"grad_norm": 0.26178088784217834,
"learning_rate": 6.296296175201564e-05,
"loss": 0.983,
"step": 2855
},
{
"epoch": 0.6598984771573604,
"grad_norm": 0.25734785199165344,
"learning_rate": 6.258913809170168e-05,
"loss": 0.9689,
"step": 2860
},
{
"epoch": 0.6610521458237194,
"grad_norm": 0.2652340829372406,
"learning_rate": 6.22159213181533e-05,
"loss": 0.9831,
"step": 2865
},
{
"epoch": 0.6622058144900784,
"grad_norm": 0.2596532702445984,
"learning_rate": 6.18433174857705e-05,
"loss": 0.9695,
"step": 2870
},
{
"epoch": 0.6633594831564374,
"grad_norm": 0.2508351504802704,
"learning_rate": 6.147133263900995e-05,
"loss": 0.9275,
"step": 2875
},
{
"epoch": 0.6645131518227965,
"grad_norm": 0.26645687222480774,
"learning_rate": 6.1099972812287e-05,
"loss": 0.9584,
"step": 2880
},
{
"epoch": 0.6656668204891555,
"grad_norm": 0.3169649541378021,
"learning_rate": 6.072924402987785e-05,
"loss": 0.9506,
"step": 2885
},
{
"epoch": 0.6668204891555145,
"grad_norm": 0.28282660245895386,
"learning_rate": 6.0359152305821766e-05,
"loss": 0.9557,
"step": 2890
},
{
"epoch": 0.6679741578218735,
"grad_norm": 0.26345473527908325,
"learning_rate": 5.99897036438235e-05,
"loss": 0.9736,
"step": 2895
},
{
"epoch": 0.6691278264882325,
"grad_norm": 0.2629302442073822,
"learning_rate": 5.962090403715592e-05,
"loss": 0.9299,
"step": 2900
},
{
"epoch": 0.6702814951545916,
"grad_norm": 0.2775762677192688,
"learning_rate": 5.925275946856275e-05,
"loss": 0.948,
"step": 2905
},
{
"epoch": 0.6714351638209506,
"grad_norm": 0.2726392149925232,
"learning_rate": 5.8885275910161576e-05,
"loss": 0.9188,
"step": 2910
},
{
"epoch": 0.6725888324873096,
"grad_norm": 0.26594123244285583,
"learning_rate": 5.8518459323346974e-05,
"loss": 0.9756,
"step": 2915
},
{
"epoch": 0.6737425011536686,
"grad_norm": 0.2796771824359894,
"learning_rate": 5.8152315658693765e-05,
"loss": 0.9523,
"step": 2920
},
{
"epoch": 0.6748961698200276,
"grad_norm": 0.2691529095172882,
"learning_rate": 5.7786850855860376e-05,
"loss": 0.9648,
"step": 2925
},
{
"epoch": 0.6760498384863867,
"grad_norm": 0.25442981719970703,
"learning_rate": 5.7422070843492734e-05,
"loss": 0.9237,
"step": 2930
},
{
"epoch": 0.6772035071527457,
"grad_norm": 0.25092652440071106,
"learning_rate": 5.7057981539127936e-05,
"loss": 0.9545,
"step": 2935
},
{
"epoch": 0.6783571758191047,
"grad_norm": 0.24751439690589905,
"learning_rate": 5.6694588849098154e-05,
"loss": 0.9359,
"step": 2940
},
{
"epoch": 0.6795108444854637,
"grad_norm": 0.25949132442474365,
"learning_rate": 5.633189866843507e-05,
"loss": 0.9119,
"step": 2945
},
{
"epoch": 0.6806645131518227,
"grad_norm": 0.2540017366409302,
"learning_rate": 5.596991688077409e-05,
"loss": 0.938,
"step": 2950
},
{
"epoch": 0.6818181818181818,
"grad_norm": 0.2602344751358032,
"learning_rate": 5.560864935825882e-05,
"loss": 0.9278,
"step": 2955
},
{
"epoch": 0.6829718504845408,
"grad_norm": 0.2632952332496643,
"learning_rate": 5.5248101961446065e-05,
"loss": 0.9745,
"step": 2960
},
{
"epoch": 0.6841255191508998,
"grad_norm": 0.25782305002212524,
"learning_rate": 5.4888280539210433e-05,
"loss": 0.9515,
"step": 2965
},
{
"epoch": 0.6852791878172588,
"grad_norm": 0.25692427158355713,
"learning_rate": 5.4529190928649754e-05,
"loss": 0.956,
"step": 2970
},
{
"epoch": 0.686432856483618,
"grad_norm": 0.2693958282470703,
"learning_rate": 5.417083895499024e-05,
"loss": 0.9965,
"step": 2975
},
{
"epoch": 0.687586525149977,
"grad_norm": 0.26253119111061096,
"learning_rate": 5.381323043149191e-05,
"loss": 0.998,
"step": 2980
},
{
"epoch": 0.688740193816336,
"grad_norm": 0.25917479395866394,
"learning_rate": 5.345637115935451e-05,
"loss": 0.9278,
"step": 2985
},
{
"epoch": 0.689893862482695,
"grad_norm": 0.25027045607566833,
"learning_rate": 5.3100266927623156e-05,
"loss": 0.9786,
"step": 2990
},
{
"epoch": 0.691047531149054,
"grad_norm": 0.25722038745880127,
"learning_rate": 5.274492351309461e-05,
"loss": 0.9346,
"step": 2995
},
{
"epoch": 0.6922011998154131,
"grad_norm": 0.2600374221801758,
"learning_rate": 5.2390346680223535e-05,
"loss": 0.9671,
"step": 3000
},
{
"epoch": 0.6933548684817721,
"grad_norm": 0.26606494188308716,
"learning_rate": 5.20365421810288e-05,
"loss": 0.9662,
"step": 3005
},
{
"epoch": 0.6945085371481311,
"grad_norm": 0.24954178929328918,
"learning_rate": 5.168351575500049e-05,
"loss": 0.9459,
"step": 3010
},
{
"epoch": 0.6956622058144901,
"grad_norm": 0.2654220759868622,
"learning_rate": 5.133127312900652e-05,
"loss": 0.9687,
"step": 3015
},
{
"epoch": 0.6968158744808491,
"grad_norm": 0.26062485575675964,
"learning_rate": 5.097982001719993e-05,
"loss": 0.9419,
"step": 3020
},
{
"epoch": 0.6979695431472082,
"grad_norm": 0.2603178322315216,
"learning_rate": 5.062916212092594e-05,
"loss": 0.9446,
"step": 3025
},
{
"epoch": 0.6991232118135672,
"grad_norm": 0.261737585067749,
"learning_rate": 5.027930512862976e-05,
"loss": 0.9457,
"step": 3030
},
{
"epoch": 0.7002768804799262,
"grad_norm": 0.2532404959201813,
"learning_rate": 4.993025471576417e-05,
"loss": 0.9526,
"step": 3035
},
{
"epoch": 0.7014305491462852,
"grad_norm": 0.2621046006679535,
"learning_rate": 4.958201654469731e-05,
"loss": 0.947,
"step": 3040
},
{
"epoch": 0.7025842178126442,
"grad_norm": 0.28589344024658203,
"learning_rate": 4.9234596264621136e-05,
"loss": 0.9693,
"step": 3045
},
{
"epoch": 0.7037378864790033,
"grad_norm": 0.2677820324897766,
"learning_rate": 4.888799951145948e-05,
"loss": 0.959,
"step": 3050
},
{
"epoch": 0.7048915551453623,
"grad_norm": 0.26197487115859985,
"learning_rate": 4.854223190777681e-05,
"loss": 0.9605,
"step": 3055
},
{
"epoch": 0.7060452238117213,
"grad_norm": 0.26514580845832825,
"learning_rate": 4.8197299062686995e-05,
"loss": 0.9486,
"step": 3060
},
{
"epoch": 0.7071988924780803,
"grad_norm": 0.25191399455070496,
"learning_rate": 4.785320657176217e-05,
"loss": 0.944,
"step": 3065
},
{
"epoch": 0.7083525611444393,
"grad_norm": 0.2851702570915222,
"learning_rate": 4.7509960016942144e-05,
"loss": 0.954,
"step": 3070
},
{
"epoch": 0.7095062298107984,
"grad_norm": 0.27994558215141296,
"learning_rate": 4.716756496644381e-05,
"loss": 0.9459,
"step": 3075
},
{
"epoch": 0.7106598984771574,
"grad_norm": 0.2702610492706299,
"learning_rate": 4.682602697467067e-05,
"loss": 0.9566,
"step": 3080
},
{
"epoch": 0.7118135671435164,
"grad_norm": 0.2700541019439697,
"learning_rate": 4.648535158212296e-05,
"loss": 0.9475,
"step": 3085
},
{
"epoch": 0.7129672358098754,
"grad_norm": 0.25302648544311523,
"learning_rate": 4.6145544315307534e-05,
"loss": 0.9379,
"step": 3090
},
{
"epoch": 0.7141209044762344,
"grad_norm": 0.25596538186073303,
"learning_rate": 4.5806610686648435e-05,
"loss": 0.9658,
"step": 3095
},
{
"epoch": 0.7152745731425935,
"grad_norm": 0.25866860151290894,
"learning_rate": 4.546855619439734e-05,
"loss": 0.9469,
"step": 3100
},
{
"epoch": 0.7164282418089525,
"grad_norm": 0.26797473430633545,
"learning_rate": 4.513138632254432e-05,
"loss": 0.9394,
"step": 3105
},
{
"epoch": 0.7175819104753115,
"grad_norm": 0.2578045129776001,
"learning_rate": 4.479510654072909e-05,
"loss": 0.9381,
"step": 3110
},
{
"epoch": 0.7187355791416705,
"grad_norm": 0.2649901211261749,
"learning_rate": 4.4459722304151965e-05,
"loss": 0.9537,
"step": 3115
},
{
"epoch": 0.7198892478080295,
"grad_norm": 0.2583473324775696,
"learning_rate": 4.412523905348568e-05,
"loss": 0.9785,
"step": 3120
},
{
"epoch": 0.7210429164743886,
"grad_norm": 0.2813000977039337,
"learning_rate": 4.379166221478697e-05,
"loss": 0.9683,
"step": 3125
},
{
"epoch": 0.7221965851407476,
"grad_norm": 0.26045987010002136,
"learning_rate": 4.345899719940843e-05,
"loss": 0.9314,
"step": 3130
},
{
"epoch": 0.7233502538071066,
"grad_norm": 0.25337517261505127,
"learning_rate": 4.312724940391114e-05,
"loss": 0.9132,
"step": 3135
},
{
"epoch": 0.7245039224734656,
"grad_norm": 0.2751719355583191,
"learning_rate": 4.279642420997655e-05,
"loss": 0.9446,
"step": 3140
},
{
"epoch": 0.7256575911398246,
"grad_norm": 0.2659566402435303,
"learning_rate": 4.246652698431969e-05,
"loss": 0.9171,
"step": 3145
},
{
"epoch": 0.7268112598061837,
"grad_norm": 0.26259174942970276,
"learning_rate": 4.213756307860175e-05,
"loss": 0.9713,
"step": 3150
},
{
"epoch": 0.7279649284725427,
"grad_norm": 0.2656116187572479,
"learning_rate": 4.180953782934351e-05,
"loss": 0.9383,
"step": 3155
},
{
"epoch": 0.7291185971389017,
"grad_norm": 0.26042497158050537,
"learning_rate": 4.148245655783869e-05,
"loss": 0.9537,
"step": 3160
},
{
"epoch": 0.7302722658052607,
"grad_norm": 0.2528819739818573,
"learning_rate": 4.115632457006746e-05,
"loss": 0.9687,
"step": 3165
},
{
"epoch": 0.7314259344716197,
"grad_norm": 0.2507370412349701,
"learning_rate": 4.0831147156610684e-05,
"loss": 0.9856,
"step": 3170
},
{
"epoch": 0.7325796031379788,
"grad_norm": 0.2561616003513336,
"learning_rate": 4.050692959256377e-05,
"loss": 0.9103,
"step": 3175
},
{
"epoch": 0.7337332718043378,
"grad_norm": 0.25249841809272766,
"learning_rate": 4.018367713745137e-05,
"loss": 0.9277,
"step": 3180
},
{
"epoch": 0.7348869404706968,
"grad_norm": 0.25682583451271057,
"learning_rate": 3.986139503514194e-05,
"loss": 0.9787,
"step": 3185
},
{
"epoch": 0.7360406091370558,
"grad_norm": 0.2522560656070709,
"learning_rate": 3.954008851376252e-05,
"loss": 0.9415,
"step": 3190
},
{
"epoch": 0.7371942778034148,
"grad_norm": 0.259802907705307,
"learning_rate": 3.9219762785614246e-05,
"loss": 0.9338,
"step": 3195
},
{
"epoch": 0.7383479464697739,
"grad_norm": 0.2639108896255493,
"learning_rate": 3.8900423047087585e-05,
"loss": 0.9653,
"step": 3200
},
{
"epoch": 0.7395016151361329,
"grad_norm": 0.27184954285621643,
"learning_rate": 3.8582074478577966e-05,
"loss": 0.9195,
"step": 3205
},
{
"epoch": 0.7406552838024919,
"grad_norm": 0.2584899961948395,
"learning_rate": 3.826472224440202e-05,
"loss": 0.9221,
"step": 3210
},
{
"epoch": 0.7418089524688509,
"grad_norm": 0.25977569818496704,
"learning_rate": 3.794837149271345e-05,
"loss": 0.9724,
"step": 3215
},
{
"epoch": 0.7429626211352099,
"grad_norm": 0.2708073556423187,
"learning_rate": 3.763302735541987e-05,
"loss": 0.9661,
"step": 3220
},
{
"epoch": 0.744116289801569,
"grad_norm": 0.2669990360736847,
"learning_rate": 3.731869494809934e-05,
"loss": 0.9382,
"step": 3225
},
{
"epoch": 0.745269958467928,
"grad_norm": 0.2620033621788025,
"learning_rate": 3.7005379369917325e-05,
"loss": 0.9837,
"step": 3230
},
{
"epoch": 0.746423627134287,
"grad_norm": 0.25972065329551697,
"learning_rate": 3.6693085703544226e-05,
"loss": 0.9258,
"step": 3235
},
{
"epoch": 0.747577295800646,
"grad_norm": 0.26290494203567505,
"learning_rate": 3.638181901507265e-05,
"loss": 0.952,
"step": 3240
},
{
"epoch": 0.748730964467005,
"grad_norm": 0.2657393217086792,
"learning_rate": 3.607158435393544e-05,
"loss": 0.9421,
"step": 3245
},
{
"epoch": 0.7498846331333641,
"grad_norm": 0.2849307954311371,
"learning_rate": 3.576238675282364e-05,
"loss": 0.9276,
"step": 3250
},
{
"epoch": 0.7510383017997231,
"grad_norm": 0.2654859721660614,
"learning_rate": 3.545423122760493e-05,
"loss": 0.9299,
"step": 3255
},
{
"epoch": 0.7521919704660821,
"grad_norm": 0.2530565559864044,
"learning_rate": 3.5147122777242204e-05,
"loss": 0.9545,
"step": 3260
},
{
"epoch": 0.7533456391324411,
"grad_norm": 0.26191383600234985,
"learning_rate": 3.4841066383712404e-05,
"loss": 0.952,
"step": 3265
},
{
"epoch": 0.7544993077988001,
"grad_norm": 0.25318285822868347,
"learning_rate": 3.4536067011925945e-05,
"loss": 0.9139,
"step": 3270
},
{
"epoch": 0.7556529764651592,
"grad_norm": 0.2692861557006836,
"learning_rate": 3.423212960964586e-05,
"loss": 0.9692,
"step": 3275
},
{
"epoch": 0.7568066451315182,
"grad_norm": 0.27093103528022766,
"learning_rate": 3.3929259107407784e-05,
"loss": 0.9167,
"step": 3280
},
{
"epoch": 0.7579603137978772,
"grad_norm": 0.25663748383522034,
"learning_rate": 3.362746041843991e-05,
"loss": 0.9418,
"step": 3285
},
{
"epoch": 0.7591139824642362,
"grad_norm": 0.2527766823768616,
"learning_rate": 3.3326738438583114e-05,
"loss": 0.9566,
"step": 3290
},
{
"epoch": 0.7602676511305952,
"grad_norm": 0.25249361991882324,
"learning_rate": 3.302709804621184e-05,
"loss": 0.9745,
"step": 3295
},
{
"epoch": 0.7614213197969543,
"grad_norm": 0.2630962133407593,
"learning_rate": 3.272854410215467e-05,
"loss": 0.944,
"step": 3300
},
{
"epoch": 0.7625749884633133,
"grad_norm": 0.2653927206993103,
"learning_rate": 3.243108144961563e-05,
"loss": 0.9386,
"step": 3305
},
{
"epoch": 0.7637286571296723,
"grad_norm": 0.25982698798179626,
"learning_rate": 3.213471491409568e-05,
"loss": 0.9665,
"step": 3310
},
{
"epoch": 0.7648823257960313,
"grad_norm": 0.24975992739200592,
"learning_rate": 3.183944930331415e-05,
"loss": 0.9397,
"step": 3315
},
{
"epoch": 0.7660359944623903,
"grad_norm": 0.26783668994903564,
"learning_rate": 3.154528940713113e-05,
"loss": 0.9349,
"step": 3320
},
{
"epoch": 0.7671896631287495,
"grad_norm": 0.26724961400032043,
"learning_rate": 3.1252239997469514e-05,
"loss": 0.9726,
"step": 3325
},
{
"epoch": 0.7683433317951085,
"grad_norm": 0.25978267192840576,
"learning_rate": 3.096030582823757e-05,
"loss": 0.963,
"step": 3330
},
{
"epoch": 0.7694970004614675,
"grad_norm": 0.2639058530330658,
"learning_rate": 3.066949163525205e-05,
"loss": 0.9452,
"step": 3335
},
{
"epoch": 0.7706506691278265,
"grad_norm": 0.25383907556533813,
"learning_rate": 3.0379802136161074e-05,
"loss": 0.9211,
"step": 3340
},
{
"epoch": 0.7718043377941856,
"grad_norm": 0.253352552652359,
"learning_rate": 3.009124203036785e-05,
"loss": 0.9334,
"step": 3345
},
{
"epoch": 0.7729580064605446,
"grad_norm": 0.25482332706451416,
"learning_rate": 2.9803815998954332e-05,
"loss": 0.9433,
"step": 3350
},
{
"epoch": 0.7741116751269036,
"grad_norm": 0.2710963189601898,
"learning_rate": 2.951752870460519e-05,
"loss": 0.9409,
"step": 3355
},
{
"epoch": 0.7752653437932626,
"grad_norm": 0.2619101107120514,
"learning_rate": 2.9232384791532375e-05,
"loss": 0.971,
"step": 3360
},
{
"epoch": 0.7764190124596216,
"grad_norm": 0.27144715189933777,
"learning_rate": 2.8948388885399568e-05,
"loss": 0.9384,
"step": 3365
},
{
"epoch": 0.7775726811259807,
"grad_norm": 0.25738927721977234,
"learning_rate": 2.8665545593247312e-05,
"loss": 0.936,
"step": 3370
},
{
"epoch": 0.7787263497923397,
"grad_norm": 0.2678247094154358,
"learning_rate": 2.838385950341821e-05,
"loss": 0.9503,
"step": 3375
},
{
"epoch": 0.7798800184586987,
"grad_norm": 0.25635334849357605,
"learning_rate": 2.810333518548246e-05,
"loss": 0.9458,
"step": 3380
},
{
"epoch": 0.7810336871250577,
"grad_norm": 0.26368802785873413,
"learning_rate": 2.7823977190163786e-05,
"loss": 0.952,
"step": 3385
},
{
"epoch": 0.7821873557914167,
"grad_norm": 0.25423574447631836,
"learning_rate": 2.754579004926551e-05,
"loss": 0.9813,
"step": 3390
},
{
"epoch": 0.7833410244577758,
"grad_norm": 0.2504977285861969,
"learning_rate": 2.7268778275597218e-05,
"loss": 0.9251,
"step": 3395
},
{
"epoch": 0.7844946931241348,
"grad_norm": 0.26885226368904114,
"learning_rate": 2.699294636290134e-05,
"loss": 0.9533,
"step": 3400
},
{
"epoch": 0.7856483617904938,
"grad_norm": 0.2703065276145935,
"learning_rate": 2.6718298785780426e-05,
"loss": 0.9293,
"step": 3405
},
{
"epoch": 0.7868020304568528,
"grad_norm": 0.25872454047203064,
"learning_rate": 2.6444839999624494e-05,
"loss": 0.9885,
"step": 3410
},
{
"epoch": 0.7879556991232118,
"grad_norm": 0.2541477680206299,
"learning_rate": 2.6172574440538678e-05,
"loss": 0.9365,
"step": 3415
},
{
"epoch": 0.7891093677895709,
"grad_norm": 0.26085081696510315,
"learning_rate": 2.5901506525271425e-05,
"loss": 0.9615,
"step": 3420
},
{
"epoch": 0.7902630364559299,
"grad_norm": 0.2733447253704071,
"learning_rate": 2.5631640651142653e-05,
"loss": 0.955,
"step": 3425
},
{
"epoch": 0.7914167051222889,
"grad_norm": 0.26850977540016174,
"learning_rate": 2.5362981195972625e-05,
"loss": 0.9904,
"step": 3430
},
{
"epoch": 0.7925703737886479,
"grad_norm": 0.25592347979545593,
"learning_rate": 2.509553251801078e-05,
"loss": 0.9888,
"step": 3435
},
{
"epoch": 0.7937240424550069,
"grad_norm": 0.25363269448280334,
"learning_rate": 2.482929895586502e-05,
"loss": 0.922,
"step": 3440
},
{
"epoch": 0.794877711121366,
"grad_norm": 0.26529207825660706,
"learning_rate": 2.4564284828431495e-05,
"loss": 0.9167,
"step": 3445
},
{
"epoch": 0.796031379787725,
"grad_norm": 0.264886736869812,
"learning_rate": 2.4300494434824373e-05,
"loss": 0.9587,
"step": 3450
},
{
"epoch": 0.797185048454084,
"grad_norm": 0.26478299498558044,
"learning_rate": 2.403793205430612e-05,
"loss": 0.9532,
"step": 3455
},
{
"epoch": 0.798338717120443,
"grad_norm": 0.26431044936180115,
"learning_rate": 2.3776601946218223e-05,
"loss": 0.9533,
"step": 3460
},
{
"epoch": 0.799492385786802,
"grad_norm": 0.2664218246936798,
"learning_rate": 2.351650834991187e-05,
"loss": 0.9451,
"step": 3465
},
{
"epoch": 0.8006460544531611,
"grad_norm": 0.25591370463371277,
"learning_rate": 2.3257655484679374e-05,
"loss": 0.9164,
"step": 3470
},
{
"epoch": 0.8017997231195201,
"grad_norm": 0.2605327069759369,
"learning_rate": 2.300004754968568e-05,
"loss": 0.9632,
"step": 3475
},
{
"epoch": 0.8029533917858791,
"grad_norm": 0.26043200492858887,
"learning_rate": 2.274368872390009e-05,
"loss": 0.9781,
"step": 3480
},
{
"epoch": 0.8041070604522381,
"grad_norm": 0.256830632686615,
"learning_rate": 2.2488583166028754e-05,
"loss": 0.9832,
"step": 3485
},
{
"epoch": 0.8052607291185971,
"grad_norm": 0.2615990936756134,
"learning_rate": 2.2234735014446907e-05,
"loss": 0.9165,
"step": 3490
},
{
"epoch": 0.8064143977849562,
"grad_norm": 0.2640646994113922,
"learning_rate": 2.1982148387131972e-05,
"loss": 0.9642,
"step": 3495
},
{
"epoch": 0.8075680664513152,
"grad_norm": 0.28312572836875916,
"learning_rate": 2.1730827381596643e-05,
"loss": 0.9532,
"step": 3500
},
{
"epoch": 0.8087217351176742,
"grad_norm": 0.2744578421115875,
"learning_rate": 2.148077607482242e-05,
"loss": 0.9206,
"step": 3505
},
{
"epoch": 0.8098754037840332,
"grad_norm": 0.266108900308609,
"learning_rate": 2.123199852319352e-05,
"loss": 0.955,
"step": 3510
},
{
"epoch": 0.8110290724503922,
"grad_norm": 0.26437053084373474,
"learning_rate": 2.098449876243096e-05,
"loss": 0.9314,
"step": 3515
},
{
"epoch": 0.8121827411167513,
"grad_norm": 0.2548779547214508,
"learning_rate": 2.0738280807527276e-05,
"loss": 0.9231,
"step": 3520
},
{
"epoch": 0.8133364097831103,
"grad_norm": 0.2563740015029907,
"learning_rate": 2.0493348652681187e-05,
"loss": 0.9955,
"step": 3525
},
{
"epoch": 0.8144900784494693,
"grad_norm": 0.25444361567497253,
"learning_rate": 2.024970627123295e-05,
"loss": 0.9399,
"step": 3530
},
{
"epoch": 0.8156437471158283,
"grad_norm": 0.2691180109977722,
"learning_rate": 2.0007357615599863e-05,
"loss": 0.937,
"step": 3535
},
{
"epoch": 0.8167974157821873,
"grad_norm": 0.2676614224910736,
"learning_rate": 1.9766306617212072e-05,
"loss": 0.9454,
"step": 3540
},
{
"epoch": 0.8179510844485464,
"grad_norm": 0.2619343101978302,
"learning_rate": 1.9526557186448922e-05,
"loss": 0.9622,
"step": 3545
},
{
"epoch": 0.8191047531149054,
"grad_norm": 0.25831881165504456,
"learning_rate": 1.9288113212575452e-05,
"loss": 0.9554,
"step": 3550
},
{
"epoch": 0.8202584217812644,
"grad_norm": 0.27389928698539734,
"learning_rate": 1.9050978563679245e-05,
"loss": 0.9845,
"step": 3555
},
{
"epoch": 0.8214120904476234,
"grad_norm": 0.2780735194683075,
"learning_rate": 1.8815157086607826e-05,
"loss": 0.9868,
"step": 3560
},
{
"epoch": 0.8225657591139824,
"grad_norm": 0.26343536376953125,
"learning_rate": 1.858065260690609e-05,
"loss": 0.9432,
"step": 3565
},
{
"epoch": 0.8237194277803415,
"grad_norm": 0.2577558159828186,
"learning_rate": 1.8347468928754407e-05,
"loss": 0.9374,
"step": 3570
},
{
"epoch": 0.8248730964467005,
"grad_norm": 0.2493932992219925,
"learning_rate": 1.811560983490682e-05,
"loss": 0.9687,
"step": 3575
},
{
"epoch": 0.8260267651130595,
"grad_norm": 0.2670118510723114,
"learning_rate": 1.78850790866296e-05,
"loss": 0.981,
"step": 3580
},
{
"epoch": 0.8271804337794185,
"grad_norm": 0.2631658613681793,
"learning_rate": 1.765588042364045e-05,
"loss": 0.9775,
"step": 3585
},
{
"epoch": 0.8283341024457775,
"grad_norm": 0.2607511878013611,
"learning_rate": 1.7428017564047594e-05,
"loss": 0.9409,
"step": 3590
},
{
"epoch": 0.8294877711121366,
"grad_norm": 0.283981591463089,
"learning_rate": 1.7201494204289647e-05,
"loss": 1.015,
"step": 3595
},
{
"epoch": 0.8306414397784956,
"grad_norm": 0.2698068916797638,
"learning_rate": 1.697631401907559e-05,
"loss": 0.9332,
"step": 3600
},
{
"epoch": 0.8317951084448546,
"grad_norm": 0.2729108929634094,
"learning_rate": 1.6752480661325078e-05,
"loss": 0.9455,
"step": 3605
},
{
"epoch": 0.8329487771112136,
"grad_norm": 0.2679850459098816,
"learning_rate": 1.6529997762109317e-05,
"loss": 0.9306,
"step": 3610
},
{
"epoch": 0.8341024457775726,
"grad_norm": 0.2747521698474884,
"learning_rate": 1.6308868930592014e-05,
"loss": 0.9063,
"step": 3615
},
{
"epoch": 0.8352561144439317,
"grad_norm": 0.26654738187789917,
"learning_rate": 1.608909775397106e-05,
"loss": 0.9431,
"step": 3620
},
{
"epoch": 0.8364097831102907,
"grad_norm": 0.2640974521636963,
"learning_rate": 1.5870687797420013e-05,
"loss": 0.9476,
"step": 3625
},
{
"epoch": 0.8375634517766497,
"grad_norm": 0.259022057056427,
"learning_rate": 1.565364260403055e-05,
"loss": 0.9037,
"step": 3630
},
{
"epoch": 0.8387171204430087,
"grad_norm": 0.2541236877441406,
"learning_rate": 1.5437965694754842e-05,
"loss": 0.9287,
"step": 3635
},
{
"epoch": 0.8398707891093677,
"grad_norm": 0.27789777517318726,
"learning_rate": 1.5223660568348442e-05,
"loss": 0.9304,
"step": 3640
},
{
"epoch": 0.8410244577757268,
"grad_norm": 0.259775847196579,
"learning_rate": 1.5010730701313625e-05,
"loss": 0.9337,
"step": 3645
},
{
"epoch": 0.8421781264420858,
"grad_norm": 0.26516568660736084,
"learning_rate": 1.4799179547842822e-05,
"loss": 0.9638,
"step": 3650
},
{
"epoch": 0.8433317951084448,
"grad_norm": 0.26108843088150024,
"learning_rate": 1.4589010539762794e-05,
"loss": 0.9627,
"step": 3655
},
{
"epoch": 0.8444854637748038,
"grad_norm": 0.2728930711746216,
"learning_rate": 1.4380227086478815e-05,
"loss": 0.9148,
"step": 3660
},
{
"epoch": 0.8456391324411628,
"grad_norm": 0.26303428411483765,
"learning_rate": 1.417283257491936e-05,
"loss": 0.9355,
"step": 3665
},
{
"epoch": 0.846792801107522,
"grad_norm": 0.25789615511894226,
"learning_rate": 1.3966830369481232e-05,
"loss": 0.9213,
"step": 3670
},
{
"epoch": 0.847946469773881,
"grad_norm": 0.25677064061164856,
"learning_rate": 1.3762223811975005e-05,
"loss": 0.9607,
"step": 3675
},
{
"epoch": 0.84910013844024,
"grad_norm": 0.25708821415901184,
"learning_rate": 1.3559016221570663e-05,
"loss": 0.9251,
"step": 3680
},
{
"epoch": 0.850253807106599,
"grad_norm": 0.26253747940063477,
"learning_rate": 1.3357210894743954e-05,
"loss": 0.9342,
"step": 3685
},
{
"epoch": 0.8514074757729581,
"grad_norm": 0.2784782946109772,
"learning_rate": 1.3156811105222721e-05,
"loss": 0.924,
"step": 3690
},
{
"epoch": 0.8525611444393171,
"grad_norm": 0.2603437304496765,
"learning_rate": 1.295782010393396e-05,
"loss": 0.9589,
"step": 3695
},
{
"epoch": 0.8537148131056761,
"grad_norm": 0.2674865126609802,
"learning_rate": 1.2760241118951011e-05,
"loss": 0.9642,
"step": 3700
},
{
"epoch": 0.8548684817720351,
"grad_norm": 0.28415897488594055,
"learning_rate": 1.256407735544114e-05,
"loss": 0.9345,
"step": 3705
},
{
"epoch": 0.8560221504383941,
"grad_norm": 0.2664172947406769,
"learning_rate": 1.2369331995613665e-05,
"loss": 0.9703,
"step": 3710
},
{
"epoch": 0.8571758191047532,
"grad_norm": 0.2628554403781891,
"learning_rate": 1.2176008198668164e-05,
"loss": 0.953,
"step": 3715
},
{
"epoch": 0.8583294877711122,
"grad_norm": 0.26892125606536865,
"learning_rate": 1.1984109100743446e-05,
"loss": 0.964,
"step": 3720
},
{
"epoch": 0.8594831564374712,
"grad_norm": 0.2509150803089142,
"learning_rate": 1.179363781486651e-05,
"loss": 0.9529,
"step": 3725
},
{
"epoch": 0.8606368251038302,
"grad_norm": 0.263091504573822,
"learning_rate": 1.160459743090203e-05,
"loss": 0.9439,
"step": 3730
},
{
"epoch": 0.8617904937701892,
"grad_norm": 0.26646509766578674,
"learning_rate": 1.1416991015502388e-05,
"loss": 0.9699,
"step": 3735
},
{
"epoch": 0.8629441624365483,
"grad_norm": 0.26804062724113464,
"learning_rate": 1.1230821612057751e-05,
"loss": 0.9448,
"step": 3740
},
{
"epoch": 0.8640978311029073,
"grad_norm": 0.2859780192375183,
"learning_rate": 1.1046092240646865e-05,
"loss": 0.939,
"step": 3745
},
{
"epoch": 0.8652514997692663,
"grad_norm": 0.257179856300354,
"learning_rate": 1.0862805897987894e-05,
"loss": 0.9542,
"step": 3750
},
{
"epoch": 0.8664051684356253,
"grad_norm": 0.26742085814476013,
"learning_rate": 1.0680965557389932e-05,
"loss": 0.9531,
"step": 3755
},
{
"epoch": 0.8675588371019843,
"grad_norm": 0.2629797160625458,
"learning_rate": 1.0500574168704746e-05,
"loss": 0.9647,
"step": 3760
},
{
"epoch": 0.8687125057683434,
"grad_norm": 0.2643450200557709,
"learning_rate": 1.0321634658278834e-05,
"loss": 0.9364,
"step": 3765
},
{
"epoch": 0.8698661744347024,
"grad_norm": 0.2526227831840515,
"learning_rate": 1.014414992890611e-05,
"loss": 0.9297,
"step": 3770
},
{
"epoch": 0.8710198431010614,
"grad_norm": 0.2794305682182312,
"learning_rate": 9.968122859780648e-06,
"loss": 0.983,
"step": 3775
},
{
"epoch": 0.8721735117674204,
"grad_norm": 0.2461376041173935,
"learning_rate": 9.793556306450125e-06,
"loss": 0.9312,
"step": 3780
},
{
"epoch": 0.8733271804337794,
"grad_norm": 0.2625856101512909,
"learning_rate": 9.6204531007694e-06,
"loss": 0.9691,
"step": 3785
},
{
"epoch": 0.8744808491001385,
"grad_norm": 0.2845028340816498,
"learning_rate": 9.44881605085456e-06,
"loss": 0.9782,
"step": 3790
},
{
"epoch": 0.8756345177664975,
"grad_norm": 0.258274644613266,
"learning_rate": 9.278647941037477e-06,
"loss": 0.9503,
"step": 3795
},
{
"epoch": 0.8767881864328565,
"grad_norm": 0.2558533847332001,
"learning_rate": 9.10995153182056e-06,
"loss": 0.9396,
"step": 3800
},
{
"epoch": 0.8779418550992155,
"grad_norm": 0.2619149684906006,
"learning_rate": 8.94272955983192e-06,
"loss": 0.9637,
"step": 3805
},
{
"epoch": 0.8790955237655745,
"grad_norm": 0.27039748430252075,
"learning_rate": 8.776984737781135e-06,
"loss": 0.9782,
"step": 3810
},
{
"epoch": 0.8802491924319336,
"grad_norm": 0.26006579399108887,
"learning_rate": 8.612719754415077e-06,
"loss": 0.9673,
"step": 3815
},
{
"epoch": 0.8814028610982926,
"grad_norm": 0.2625749111175537,
"learning_rate": 8.449937274474396e-06,
"loss": 0.9224,
"step": 3820
},
{
"epoch": 0.8825565297646516,
"grad_norm": 0.26278451085090637,
"learning_rate": 8.28863993865029e-06,
"loss": 0.9762,
"step": 3825
},
{
"epoch": 0.8837101984310106,
"grad_norm": 0.25975650548934937,
"learning_rate": 8.128830363541574e-06,
"loss": 0.9697,
"step": 3830
},
{
"epoch": 0.8848638670973696,
"grad_norm": 0.2637532949447632,
"learning_rate": 7.970511141612392e-06,
"loss": 0.9278,
"step": 3835
},
{
"epoch": 0.8860175357637287,
"grad_norm": 0.2644905149936676,
"learning_rate": 7.81368484114996e-06,
"loss": 0.9218,
"step": 3840
},
{
"epoch": 0.8871712044300877,
"grad_norm": 0.25665807723999023,
"learning_rate": 7.65835400622309e-06,
"loss": 0.9872,
"step": 3845
},
{
"epoch": 0.8883248730964467,
"grad_norm": 0.26085856556892395,
"learning_rate": 7.504521156640853e-06,
"loss": 0.9236,
"step": 3850
},
{
"epoch": 0.8894785417628057,
"grad_norm": 0.2720387578010559,
"learning_rate": 7.352188787911618e-06,
"loss": 0.9589,
"step": 3855
},
{
"epoch": 0.8906322104291647,
"grad_norm": 0.26417550444602966,
"learning_rate": 7.201359371202699e-06,
"loss": 0.9664,
"step": 3860
},
{
"epoch": 0.8917858790955238,
"grad_norm": 0.2652938663959503,
"learning_rate": 7.05203535330019e-06,
"loss": 0.926,
"step": 3865
},
{
"epoch": 0.8929395477618828,
"grad_norm": 0.2653699517250061,
"learning_rate": 6.904219156569325e-06,
"loss": 0.9923,
"step": 3870
},
{
"epoch": 0.8940932164282418,
"grad_norm": 0.2711726129055023,
"learning_rate": 6.757913178915087e-06,
"loss": 0.911,
"step": 3875
},
{
"epoch": 0.8952468850946008,
"grad_norm": 0.2669640779495239,
"learning_rate": 6.613119793743428e-06,
"loss": 0.9675,
"step": 3880
},
{
"epoch": 0.8964005537609598,
"grad_norm": 0.29151153564453125,
"learning_rate": 6.469841349922689e-06,
"loss": 0.9209,
"step": 3885
},
{
"epoch": 0.8975542224273189,
"grad_norm": 0.2601694166660309,
"learning_rate": 6.32808017174551e-06,
"loss": 0.9558,
"step": 3890
},
{
"epoch": 0.8987078910936779,
"grad_norm": 0.26199638843536377,
"learning_rate": 6.187838558891157e-06,
"loss": 0.9613,
"step": 3895
},
{
"epoch": 0.8998615597600369,
"grad_norm": 0.2727982699871063,
"learning_rate": 6.049118786388152e-06,
"loss": 0.9568,
"step": 3900
},
{
"epoch": 0.9010152284263959,
"grad_norm": 0.2732504606246948,
"learning_rate": 5.911923104577455e-06,
"loss": 0.9737,
"step": 3905
},
{
"epoch": 0.9021688970927549,
"grad_norm": 0.2629091441631317,
"learning_rate": 5.7762537390758875e-06,
"loss": 0.9901,
"step": 3910
},
{
"epoch": 0.903322565759114,
"grad_norm": 0.2660592496395111,
"learning_rate": 5.642112890740036e-06,
"loss": 0.9522,
"step": 3915
},
{
"epoch": 0.904476234425473,
"grad_norm": 0.2575923502445221,
"learning_rate": 5.509502735630601e-06,
"loss": 0.9218,
"step": 3920
},
{
"epoch": 0.905629903091832,
"grad_norm": 0.25905632972717285,
"learning_rate": 5.3784254249770296e-06,
"loss": 0.9516,
"step": 3925
},
{
"epoch": 0.906783571758191,
"grad_norm": 0.29798364639282227,
"learning_rate": 5.248883085142653e-06,
"loss": 0.9691,
"step": 3930
},
{
"epoch": 0.90793724042455,
"grad_norm": 0.2637086510658264,
"learning_rate": 5.120877817590197e-06,
"loss": 0.9274,
"step": 3935
},
{
"epoch": 0.9090909090909091,
"grad_norm": 0.25364989042282104,
"learning_rate": 4.994411698847667e-06,
"loss": 0.933,
"step": 3940
},
{
"epoch": 0.9102445777572681,
"grad_norm": 0.24825413525104523,
"learning_rate": 4.869486780474685e-06,
"loss": 0.9442,
"step": 3945
},
{
"epoch": 0.9113982464236271,
"grad_norm": 0.25310808420181274,
"learning_rate": 4.746105089029229e-06,
"loss": 0.9469,
"step": 3950
},
{
"epoch": 0.9125519150899861,
"grad_norm": 0.26140066981315613,
"learning_rate": 4.624268626034667e-06,
"loss": 0.9754,
"step": 3955
},
{
"epoch": 0.9137055837563451,
"grad_norm": 0.2598501443862915,
"learning_rate": 4.50397936794742e-06,
"loss": 0.9462,
"step": 3960
},
{
"epoch": 0.9148592524227042,
"grad_norm": 0.25906556844711304,
"learning_rate": 4.385239266124752e-06,
"loss": 0.9194,
"step": 3965
},
{
"epoch": 0.9160129210890632,
"grad_norm": 0.26104477047920227,
"learning_rate": 4.268050246793276e-06,
"loss": 0.934,
"step": 3970
},
{
"epoch": 0.9171665897554222,
"grad_norm": 0.2576802968978882,
"learning_rate": 4.1524142110175745e-06,
"loss": 0.9832,
"step": 3975
},
{
"epoch": 0.9183202584217812,
"grad_norm": 0.28592053055763245,
"learning_rate": 4.038333034669406e-06,
"loss": 1.0201,
"step": 3980
},
{
"epoch": 0.9194739270881402,
"grad_norm": 0.2563294470310211,
"learning_rate": 3.9258085683972935e-06,
"loss": 0.9904,
"step": 3985
},
{
"epoch": 0.9206275957544993,
"grad_norm": 0.2748868763446808,
"learning_rate": 3.814842637596483e-06,
"loss": 0.9739,
"step": 3990
},
{
"epoch": 0.9217812644208583,
"grad_norm": 0.2553744316101074,
"learning_rate": 3.705437042379334e-06,
"loss": 1.0174,
"step": 3995
},
{
"epoch": 0.9229349330872173,
"grad_norm": 0.2620948553085327,
"learning_rate": 3.5975935575461083e-06,
"loss": 0.9456,
"step": 4000
},
{
"epoch": 0.9240886017535763,
"grad_norm": 0.2571701407432556,
"learning_rate": 3.491313932556206e-06,
"loss": 0.9339,
"step": 4005
},
{
"epoch": 0.9252422704199353,
"grad_norm": 0.26273953914642334,
"learning_rate": 3.3865998914997643e-06,
"loss": 0.9831,
"step": 4010
},
{
"epoch": 0.9263959390862944,
"grad_norm": 0.25073423981666565,
"learning_rate": 3.2834531330696694e-06,
"loss": 0.9478,
"step": 4015
},
{
"epoch": 0.9275496077526535,
"grad_norm": 0.2635844051837921,
"learning_rate": 3.1818753305340565e-06,
"loss": 0.9356,
"step": 4020
},
{
"epoch": 0.9287032764190125,
"grad_norm": 0.25664207339286804,
"learning_rate": 3.081868131709109e-06,
"loss": 0.9419,
"step": 4025
},
{
"epoch": 0.9298569450853715,
"grad_norm": 0.2655963897705078,
"learning_rate": 2.9834331589323693e-06,
"loss": 0.9114,
"step": 4030
},
{
"epoch": 0.9310106137517306,
"grad_norm": 0.2534726858139038,
"learning_rate": 2.8865720090364034e-06,
"loss": 0.9644,
"step": 4035
},
{
"epoch": 0.9321642824180896,
"grad_norm": 0.2585753798484802,
"learning_rate": 2.7912862533228558e-06,
"loss": 0.9495,
"step": 4040
},
{
"epoch": 0.9333179510844486,
"grad_norm": 0.25545534491539,
"learning_rate": 2.6975774375370464e-06,
"loss": 0.9604,
"step": 4045
},
{
"epoch": 0.9344716197508076,
"grad_norm": 0.2509816884994507,
"learning_rate": 2.6054470818428377e-06,
"loss": 0.9603,
"step": 4050
},
{
"epoch": 0.9356252884171666,
"grad_norm": 0.26575180888175964,
"learning_rate": 2.5148966807979733e-06,
"loss": 0.9527,
"step": 4055
},
{
"epoch": 0.9367789570835257,
"grad_norm": 0.25556278228759766,
"learning_rate": 2.4259277033298555e-06,
"loss": 0.9592,
"step": 4060
},
{
"epoch": 0.9379326257498847,
"grad_norm": 0.26294147968292236,
"learning_rate": 2.338541592711696e-06,
"loss": 0.954,
"step": 4065
},
{
"epoch": 0.9390862944162437,
"grad_norm": 0.25935277342796326,
"learning_rate": 2.2527397665391027e-06,
"loss": 0.9491,
"step": 4070
},
{
"epoch": 0.9402399630826027,
"grad_norm": 0.2684488594532013,
"learning_rate": 2.168523616707141e-06,
"loss": 0.9403,
"step": 4075
},
{
"epoch": 0.9413936317489617,
"grad_norm": 0.26326921582221985,
"learning_rate": 2.0858945093876316e-06,
"loss": 0.9381,
"step": 4080
},
{
"epoch": 0.9425473004153208,
"grad_norm": 0.2741606831550598,
"learning_rate": 2.0048537850071325e-06,
"loss": 0.9475,
"step": 4085
},
{
"epoch": 0.9437009690816798,
"grad_norm": 0.2589118480682373,
"learning_rate": 1.925402758225059e-06,
"loss": 0.981,
"step": 4090
},
{
"epoch": 0.9448546377480388,
"grad_norm": 0.2977379262447357,
"learning_rate": 1.8475427179124871e-06,
"loss": 0.9404,
"step": 4095
},
{
"epoch": 0.9460083064143978,
"grad_norm": 0.26037946343421936,
"learning_rate": 1.771274927131139e-06,
"loss": 0.9337,
"step": 4100
},
{
"epoch": 0.9471619750807568,
"grad_norm": 0.25153833627700806,
"learning_rate": 1.6966006231129317e-06,
"loss": 0.9531,
"step": 4105
},
{
"epoch": 0.9483156437471159,
"grad_norm": 0.25663888454437256,
"learning_rate": 1.6235210172399372e-06,
"loss": 0.972,
"step": 4110
},
{
"epoch": 0.9494693124134749,
"grad_norm": 0.24960479140281677,
"learning_rate": 1.5520372950246887e-06,
"loss": 0.9229,
"step": 4115
},
{
"epoch": 0.9506229810798339,
"grad_norm": 0.2711315453052521,
"learning_rate": 1.4821506160909493e-06,
"loss": 0.9771,
"step": 4120
},
{
"epoch": 0.9517766497461929,
"grad_norm": 0.24772925674915314,
"learning_rate": 1.4138621141549402e-06,
"loss": 0.9495,
"step": 4125
},
{
"epoch": 0.9529303184125519,
"grad_norm": 0.2522965967655182,
"learning_rate": 1.3471728970068987e-06,
"loss": 0.932,
"step": 4130
},
{
"epoch": 0.954083987078911,
"grad_norm": 0.24729067087173462,
"learning_rate": 1.2820840464931704e-06,
"loss": 0.9023,
"step": 4135
},
{
"epoch": 0.95523765574527,
"grad_norm": 0.24839915335178375,
"learning_rate": 1.2185966184985685e-06,
"loss": 0.9072,
"step": 4140
},
{
"epoch": 0.956391324411629,
"grad_norm": 0.2570428252220154,
"learning_rate": 1.1567116429293423e-06,
"loss": 0.9276,
"step": 4145
},
{
"epoch": 0.957544993077988,
"grad_norm": 0.26178133487701416,
"learning_rate": 1.0964301236963904e-06,
"loss": 0.9461,
"step": 4150
},
{
"epoch": 0.958698661744347,
"grad_norm": 0.2655417323112488,
"learning_rate": 1.0377530386990408e-06,
"loss": 0.9586,
"step": 4155
},
{
"epoch": 0.9598523304107061,
"grad_norm": 0.2616187036037445,
"learning_rate": 9.80681339809142e-07,
"loss": 0.952,
"step": 4160
},
{
"epoch": 0.9610059990770651,
"grad_norm": 0.2501654326915741,
"learning_rate": 9.252159528556403e-07,
"loss": 0.9187,
"step": 4165
},
{
"epoch": 0.9621596677434241,
"grad_norm": 0.261311411857605,
"learning_rate": 8.713577776095494e-07,
"loss": 0.9637,
"step": 4170
},
{
"epoch": 0.9633133364097831,
"grad_norm": 0.26594147086143494,
"learning_rate": 8.191076877693604e-07,
"loss": 0.946,
"step": 4175
},
{
"epoch": 0.9644670050761421,
"grad_norm": 0.25597435235977173,
"learning_rate": 7.684665309468875e-07,
"loss": 0.9229,
"step": 4180
},
{
"epoch": 0.9656206737425012,
"grad_norm": 0.2549925446510315,
"learning_rate": 7.194351286534784e-07,
"loss": 0.966,
"step": 4185
},
{
"epoch": 0.9667743424088602,
"grad_norm": 0.24757057428359985,
"learning_rate": 6.720142762867032e-07,
"loss": 0.9443,
"step": 4190
},
{
"epoch": 0.9679280110752192,
"grad_norm": 0.26646557450294495,
"learning_rate": 6.262047431174866e-07,
"loss": 0.9548,
"step": 4195
},
{
"epoch": 0.9690816797415782,
"grad_norm": 0.25552335381507874,
"learning_rate": 5.820072722775849e-07,
"loss": 0.941,
"step": 4200
},
{
"epoch": 0.9702353484079372,
"grad_norm": 0.2553008496761322,
"learning_rate": 5.394225807475284e-07,
"loss": 0.9398,
"step": 4205
},
{
"epoch": 0.9713890170742963,
"grad_norm": 0.26585790514945984,
"learning_rate": 4.984513593450424e-07,
"loss": 0.9469,
"step": 4210
},
{
"epoch": 0.9725426857406553,
"grad_norm": 0.25513309240341187,
"learning_rate": 4.5909427271374485e-07,
"loss": 0.9518,
"step": 4215
},
{
"epoch": 0.9736963544070143,
"grad_norm": 0.2652360796928406,
"learning_rate": 4.2135195931249926e-07,
"loss": 0.9489,
"step": 4220
},
{
"epoch": 0.9748500230733733,
"grad_norm": 0.2571764588356018,
"learning_rate": 3.8522503140493436e-07,
"loss": 0.9149,
"step": 4225
},
{
"epoch": 0.9760036917397323,
"grad_norm": 0.2566836178302765,
"learning_rate": 3.50714075049563e-07,
"loss": 0.9953,
"step": 4230
},
{
"epoch": 0.9771573604060914,
"grad_norm": 0.290528804063797,
"learning_rate": 3.178196500903008e-07,
"loss": 0.9222,
"step": 4235
},
{
"epoch": 0.9783110290724504,
"grad_norm": 0.2592345178127289,
"learning_rate": 2.8654229014730694e-07,
"loss": 0.9206,
"step": 4240
},
{
"epoch": 0.9794646977388094,
"grad_norm": 0.2652153968811035,
"learning_rate": 2.568825026084354e-07,
"loss": 0.9443,
"step": 4245
},
{
"epoch": 0.9806183664051684,
"grad_norm": 0.24910645186901093,
"learning_rate": 2.288407686208971e-07,
"loss": 0.9519,
"step": 4250
},
{
"epoch": 0.9817720350715274,
"grad_norm": 0.25289344787597656,
"learning_rate": 2.024175430835329e-07,
"loss": 0.9381,
"step": 4255
},
{
"epoch": 0.9829257037378865,
"grad_norm": 0.25260549783706665,
"learning_rate": 1.7761325463937494e-07,
"loss": 0.9563,
"step": 4260
},
{
"epoch": 0.9840793724042455,
"grad_norm": 0.2648310363292694,
"learning_rate": 1.5442830566874123e-07,
"loss": 0.9439,
"step": 4265
},
{
"epoch": 0.9852330410706045,
"grad_norm": 0.2567463517189026,
"learning_rate": 1.3286307228269623e-07,
"loss": 0.9289,
"step": 4270
},
{
"epoch": 0.9863867097369635,
"grad_norm": 0.2639390826225281,
"learning_rate": 1.1291790431692262e-07,
"loss": 0.9529,
"step": 4275
},
{
"epoch": 0.9875403784033225,
"grad_norm": 0.2619445323944092,
"learning_rate": 9.459312532608122e-08,
"loss": 0.9555,
"step": 4280
},
{
"epoch": 0.9886940470696816,
"grad_norm": 0.2608962059020996,
"learning_rate": 7.788903257852643e-08,
"loss": 0.9886,
"step": 4285
},
{
"epoch": 0.9898477157360406,
"grad_norm": 0.24952927231788635,
"learning_rate": 6.280589705153217e-08,
"loss": 0.9446,
"step": 4290
},
{
"epoch": 0.9910013844023996,
"grad_norm": 0.2512921988964081,
"learning_rate": 4.934396342684e-08,
"loss": 0.9613,
"step": 4295
},
{
"epoch": 0.9921550530687586,
"grad_norm": 0.2717246413230896,
"learning_rate": 3.750345008675105e-08,
"loss": 0.9526,
"step": 4300
},
{
"epoch": 0.9933087217351176,
"grad_norm": 0.2628538906574249,
"learning_rate": 2.728454911050671e-08,
"loss": 0.9467,
"step": 4305
},
{
"epoch": 0.9944623904014767,
"grad_norm": 0.26591312885284424,
"learning_rate": 1.8687426271246645e-08,
"loss": 0.9604,
"step": 4310
},
{
"epoch": 0.9956160590678357,
"grad_norm": 0.26221415400505066,
"learning_rate": 1.1712221033288728e-08,
"loss": 0.9664,
"step": 4315
},
{
"epoch": 0.9967697277341947,
"grad_norm": 0.2693996727466583,
"learning_rate": 6.359046549864189e-09,
"loss": 0.966,
"step": 4320
},
{
"epoch": 0.9979233964005537,
"grad_norm": 0.26124852895736694,
"learning_rate": 2.627989661252439e-09,
"loss": 0.9511,
"step": 4325
},
{
"epoch": 0.9990770650669127,
"grad_norm": 0.2654908299446106,
"learning_rate": 5.191108934710087e-10,
"loss": 0.9212,
"step": 4330
},
{
"epoch": 1.0,
"eval_loss": 0.9544818997383118,
"eval_runtime": 1126.36,
"eval_samples_per_second": 13.628,
"eval_steps_per_second": 0.426,
"step": 4334
},
{
"epoch": 1.0,
"step": 4334,
"total_flos": 1.2189453453533118e+19,
"train_loss": 0.9622706794551732,
"train_runtime": 38294.8783,
"train_samples_per_second": 3.622,
"train_steps_per_second": 0.113
}
],
"logging_steps": 5,
"max_steps": 4334,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.2189453453533118e+19,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}