xp185 / checkpoint-1320 /trainer_state.json
ugaoo's picture
Upload folder using huggingface_hub
2d83c46 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.975794251134644,
"eval_steps": 500,
"global_step": 1320,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0045385779122541605,
"grad_norm": 12.459190368652344,
"learning_rate": 5.0000000000000004e-08,
"loss": 4.6369,
"step": 1
},
{
"epoch": 0.009077155824508321,
"grad_norm": 12.756917953491211,
"learning_rate": 1.0000000000000001e-07,
"loss": 4.6738,
"step": 2
},
{
"epoch": 0.01361573373676248,
"grad_norm": 15.50593090057373,
"learning_rate": 1.5000000000000002e-07,
"loss": 4.6681,
"step": 3
},
{
"epoch": 0.018154311649016642,
"grad_norm": 12.976861000061035,
"learning_rate": 2.0000000000000002e-07,
"loss": 4.246,
"step": 4
},
{
"epoch": 0.0226928895612708,
"grad_norm": 14.594775199890137,
"learning_rate": 2.5000000000000004e-07,
"loss": 4.4452,
"step": 5
},
{
"epoch": 0.02723146747352496,
"grad_norm": 16.087888717651367,
"learning_rate": 3.0000000000000004e-07,
"loss": 4.5528,
"step": 6
},
{
"epoch": 0.03177004538577912,
"grad_norm": 13.62125301361084,
"learning_rate": 3.5000000000000004e-07,
"loss": 4.5628,
"step": 7
},
{
"epoch": 0.036308623298033284,
"grad_norm": 17.010616302490234,
"learning_rate": 4.0000000000000003e-07,
"loss": 4.5768,
"step": 8
},
{
"epoch": 0.04084720121028744,
"grad_norm": 14.567586898803711,
"learning_rate": 4.5000000000000003e-07,
"loss": 4.3937,
"step": 9
},
{
"epoch": 0.0453857791225416,
"grad_norm": 15.238479614257812,
"learning_rate": 5.000000000000001e-07,
"loss": 4.495,
"step": 10
},
{
"epoch": 0.049924357034795766,
"grad_norm": 13.89387321472168,
"learning_rate": 5.5e-07,
"loss": 4.4705,
"step": 11
},
{
"epoch": 0.05446293494704992,
"grad_norm": 13.949710845947266,
"learning_rate": 6.000000000000001e-07,
"loss": 4.328,
"step": 12
},
{
"epoch": 0.059001512859304085,
"grad_norm": 13.15122127532959,
"learning_rate": 6.5e-07,
"loss": 4.4296,
"step": 13
},
{
"epoch": 0.06354009077155824,
"grad_norm": 14.076905250549316,
"learning_rate": 7.000000000000001e-07,
"loss": 4.5285,
"step": 14
},
{
"epoch": 0.0680786686838124,
"grad_norm": 14.299891471862793,
"learning_rate": 7.5e-07,
"loss": 4.5849,
"step": 15
},
{
"epoch": 0.07261724659606657,
"grad_norm": 13.573644638061523,
"learning_rate": 8.000000000000001e-07,
"loss": 4.3638,
"step": 16
},
{
"epoch": 0.07715582450832073,
"grad_norm": 14.140484809875488,
"learning_rate": 8.500000000000001e-07,
"loss": 4.4449,
"step": 17
},
{
"epoch": 0.08169440242057488,
"grad_norm": 11.941351890563965,
"learning_rate": 9.000000000000001e-07,
"loss": 4.3646,
"step": 18
},
{
"epoch": 0.08623298033282904,
"grad_norm": 10.630327224731445,
"learning_rate": 9.500000000000001e-07,
"loss": 4.2902,
"step": 19
},
{
"epoch": 0.0907715582450832,
"grad_norm": 11.662637710571289,
"learning_rate": 1.0000000000000002e-06,
"loss": 4.1961,
"step": 20
},
{
"epoch": 0.09531013615733737,
"grad_norm": 12.008113861083984,
"learning_rate": 1.0500000000000001e-06,
"loss": 4.1941,
"step": 21
},
{
"epoch": 0.09984871406959153,
"grad_norm": 12.686023712158203,
"learning_rate": 1.1e-06,
"loss": 4.1346,
"step": 22
},
{
"epoch": 0.1043872919818457,
"grad_norm": 12.529243469238281,
"learning_rate": 1.1500000000000002e-06,
"loss": 4.1137,
"step": 23
},
{
"epoch": 0.10892586989409984,
"grad_norm": 11.19096565246582,
"learning_rate": 1.2000000000000002e-06,
"loss": 4.0772,
"step": 24
},
{
"epoch": 0.11346444780635401,
"grad_norm": 11.182024955749512,
"learning_rate": 1.25e-06,
"loss": 3.8888,
"step": 25
},
{
"epoch": 0.11800302571860817,
"grad_norm": 9.899381637573242,
"learning_rate": 1.3e-06,
"loss": 4.0117,
"step": 26
},
{
"epoch": 0.12254160363086233,
"grad_norm": 9.448798179626465,
"learning_rate": 1.3500000000000002e-06,
"loss": 3.8384,
"step": 27
},
{
"epoch": 0.12708018154311648,
"grad_norm": 10.632583618164062,
"learning_rate": 1.4000000000000001e-06,
"loss": 3.7857,
"step": 28
},
{
"epoch": 0.13161875945537066,
"grad_norm": 10.56238079071045,
"learning_rate": 1.45e-06,
"loss": 3.4783,
"step": 29
},
{
"epoch": 0.1361573373676248,
"grad_norm": 12.410117149353027,
"learning_rate": 1.5e-06,
"loss": 3.4828,
"step": 30
},
{
"epoch": 0.14069591527987896,
"grad_norm": 10.183599472045898,
"learning_rate": 1.5500000000000002e-06,
"loss": 3.6337,
"step": 31
},
{
"epoch": 0.14523449319213314,
"grad_norm": 9.421585083007812,
"learning_rate": 1.6000000000000001e-06,
"loss": 3.4453,
"step": 32
},
{
"epoch": 0.14977307110438728,
"grad_norm": 9.230025291442871,
"learning_rate": 1.6500000000000003e-06,
"loss": 3.3481,
"step": 33
},
{
"epoch": 0.15431164901664146,
"grad_norm": 8.295567512512207,
"learning_rate": 1.7000000000000002e-06,
"loss": 3.3145,
"step": 34
},
{
"epoch": 0.1588502269288956,
"grad_norm": 9.138203620910645,
"learning_rate": 1.75e-06,
"loss": 3.1872,
"step": 35
},
{
"epoch": 0.16338880484114976,
"grad_norm": 11.864872932434082,
"learning_rate": 1.8000000000000001e-06,
"loss": 2.9837,
"step": 36
},
{
"epoch": 0.16792738275340394,
"grad_norm": 12.373150825500488,
"learning_rate": 1.85e-06,
"loss": 3.0049,
"step": 37
},
{
"epoch": 0.17246596066565809,
"grad_norm": 21.665483474731445,
"learning_rate": 1.9000000000000002e-06,
"loss": 2.7664,
"step": 38
},
{
"epoch": 0.17700453857791226,
"grad_norm": 22.663740158081055,
"learning_rate": 1.9500000000000004e-06,
"loss": 2.7272,
"step": 39
},
{
"epoch": 0.1815431164901664,
"grad_norm": 26.674400329589844,
"learning_rate": 2.0000000000000003e-06,
"loss": 2.8077,
"step": 40
},
{
"epoch": 0.18608169440242056,
"grad_norm": 28.503612518310547,
"learning_rate": 2.05e-06,
"loss": 2.7306,
"step": 41
},
{
"epoch": 0.19062027231467474,
"grad_norm": 29.03409767150879,
"learning_rate": 2.1000000000000002e-06,
"loss": 2.7405,
"step": 42
},
{
"epoch": 0.1951588502269289,
"grad_norm": 17.844894409179688,
"learning_rate": 2.15e-06,
"loss": 2.6464,
"step": 43
},
{
"epoch": 0.19969742813918306,
"grad_norm": 19.220829010009766,
"learning_rate": 2.2e-06,
"loss": 2.4554,
"step": 44
},
{
"epoch": 0.2042360060514372,
"grad_norm": 25.161415100097656,
"learning_rate": 2.25e-06,
"loss": 2.2773,
"step": 45
},
{
"epoch": 0.2087745839636914,
"grad_norm": 17.46828269958496,
"learning_rate": 2.3000000000000004e-06,
"loss": 2.3392,
"step": 46
},
{
"epoch": 0.21331316187594554,
"grad_norm": 13.194672584533691,
"learning_rate": 2.35e-06,
"loss": 2.0409,
"step": 47
},
{
"epoch": 0.2178517397881997,
"grad_norm": 7.290500164031982,
"learning_rate": 2.4000000000000003e-06,
"loss": 2.1879,
"step": 48
},
{
"epoch": 0.22239031770045387,
"grad_norm": 7.56943941116333,
"learning_rate": 2.4500000000000003e-06,
"loss": 2.1656,
"step": 49
},
{
"epoch": 0.22692889561270801,
"grad_norm": 8.335527420043945,
"learning_rate": 2.5e-06,
"loss": 2.1294,
"step": 50
},
{
"epoch": 0.2314674735249622,
"grad_norm": 10.013853073120117,
"learning_rate": 2.55e-06,
"loss": 2.1816,
"step": 51
},
{
"epoch": 0.23600605143721634,
"grad_norm": 8.674482345581055,
"learning_rate": 2.6e-06,
"loss": 2.0078,
"step": 52
},
{
"epoch": 0.2405446293494705,
"grad_norm": 6.09174919128418,
"learning_rate": 2.6500000000000005e-06,
"loss": 2.1328,
"step": 53
},
{
"epoch": 0.24508320726172467,
"grad_norm": 7.3330488204956055,
"learning_rate": 2.7000000000000004e-06,
"loss": 1.8961,
"step": 54
},
{
"epoch": 0.24962178517397882,
"grad_norm": 8.764911651611328,
"learning_rate": 2.7500000000000004e-06,
"loss": 1.7026,
"step": 55
},
{
"epoch": 0.25416036308623297,
"grad_norm": 12.413115501403809,
"learning_rate": 2.8000000000000003e-06,
"loss": 1.8172,
"step": 56
},
{
"epoch": 0.2586989409984871,
"grad_norm": 9.009276390075684,
"learning_rate": 2.85e-06,
"loss": 1.6527,
"step": 57
},
{
"epoch": 0.2632375189107413,
"grad_norm": 9.965579986572266,
"learning_rate": 2.9e-06,
"loss": 1.6348,
"step": 58
},
{
"epoch": 0.26777609682299547,
"grad_norm": 7.091963768005371,
"learning_rate": 2.95e-06,
"loss": 1.7016,
"step": 59
},
{
"epoch": 0.2723146747352496,
"grad_norm": 5.276648998260498,
"learning_rate": 3e-06,
"loss": 1.7088,
"step": 60
},
{
"epoch": 0.27685325264750377,
"grad_norm": 9.634596824645996,
"learning_rate": 3.05e-06,
"loss": 1.4961,
"step": 61
},
{
"epoch": 0.2813918305597579,
"grad_norm": 6.663750648498535,
"learning_rate": 3.1000000000000004e-06,
"loss": 1.522,
"step": 62
},
{
"epoch": 0.2859304084720121,
"grad_norm": 36.79426956176758,
"learning_rate": 3.1500000000000003e-06,
"loss": 1.4855,
"step": 63
},
{
"epoch": 0.29046898638426627,
"grad_norm": 31.88567352294922,
"learning_rate": 3.2000000000000003e-06,
"loss": 1.5058,
"step": 64
},
{
"epoch": 0.2950075642965204,
"grad_norm": 12.319960594177246,
"learning_rate": 3.2500000000000002e-06,
"loss": 1.4547,
"step": 65
},
{
"epoch": 0.29954614220877457,
"grad_norm": 4.246046543121338,
"learning_rate": 3.3000000000000006e-06,
"loss": 1.3153,
"step": 66
},
{
"epoch": 0.3040847201210287,
"grad_norm": 5.3949503898620605,
"learning_rate": 3.3500000000000005e-06,
"loss": 1.4237,
"step": 67
},
{
"epoch": 0.3086232980332829,
"grad_norm": 9.975737571716309,
"learning_rate": 3.4000000000000005e-06,
"loss": 1.327,
"step": 68
},
{
"epoch": 0.31316187594553707,
"grad_norm": 6.356144428253174,
"learning_rate": 3.45e-06,
"loss": 1.2685,
"step": 69
},
{
"epoch": 0.3177004538577912,
"grad_norm": 3.9029836654663086,
"learning_rate": 3.5e-06,
"loss": 1.3856,
"step": 70
},
{
"epoch": 0.32223903177004537,
"grad_norm": 4.215930938720703,
"learning_rate": 3.5500000000000003e-06,
"loss": 1.1844,
"step": 71
},
{
"epoch": 0.3267776096822995,
"grad_norm": 27.889633178710938,
"learning_rate": 3.6000000000000003e-06,
"loss": 1.2535,
"step": 72
},
{
"epoch": 0.3313161875945537,
"grad_norm": 23.77273178100586,
"learning_rate": 3.65e-06,
"loss": 1.3189,
"step": 73
},
{
"epoch": 0.3358547655068079,
"grad_norm": 9.203927040100098,
"learning_rate": 3.7e-06,
"loss": 1.1993,
"step": 74
},
{
"epoch": 0.340393343419062,
"grad_norm": 3.1091806888580322,
"learning_rate": 3.7500000000000005e-06,
"loss": 1.1804,
"step": 75
},
{
"epoch": 0.34493192133131617,
"grad_norm": 9.0440034866333,
"learning_rate": 3.8000000000000005e-06,
"loss": 1.2318,
"step": 76
},
{
"epoch": 0.3494704992435703,
"grad_norm": 7.17051362991333,
"learning_rate": 3.85e-06,
"loss": 1.2018,
"step": 77
},
{
"epoch": 0.3540090771558245,
"grad_norm": 8.164457321166992,
"learning_rate": 3.900000000000001e-06,
"loss": 1.0911,
"step": 78
},
{
"epoch": 0.3585476550680787,
"grad_norm": 3.3671042919158936,
"learning_rate": 3.95e-06,
"loss": 1.1747,
"step": 79
},
{
"epoch": 0.3630862329803328,
"grad_norm": 4.223191261291504,
"learning_rate": 4.000000000000001e-06,
"loss": 1.1141,
"step": 80
},
{
"epoch": 0.367624810892587,
"grad_norm": 10.657241821289062,
"learning_rate": 4.05e-06,
"loss": 1.0506,
"step": 81
},
{
"epoch": 0.3721633888048411,
"grad_norm": 2.664783239364624,
"learning_rate": 4.1e-06,
"loss": 1.1271,
"step": 82
},
{
"epoch": 0.3767019667170953,
"grad_norm": 5.334985733032227,
"learning_rate": 4.15e-06,
"loss": 1.0812,
"step": 83
},
{
"epoch": 0.3812405446293495,
"grad_norm": 7.471070289611816,
"learning_rate": 4.2000000000000004e-06,
"loss": 1.1027,
"step": 84
},
{
"epoch": 0.3857791225416036,
"grad_norm": 3.270421028137207,
"learning_rate": 4.25e-06,
"loss": 1.054,
"step": 85
},
{
"epoch": 0.3903177004538578,
"grad_norm": 1.9382085800170898,
"learning_rate": 4.3e-06,
"loss": 0.975,
"step": 86
},
{
"epoch": 0.394856278366112,
"grad_norm": 24.2485408782959,
"learning_rate": 4.350000000000001e-06,
"loss": 1.0288,
"step": 87
},
{
"epoch": 0.39939485627836613,
"grad_norm": 42.359432220458984,
"learning_rate": 4.4e-06,
"loss": 1.111,
"step": 88
},
{
"epoch": 0.4039334341906203,
"grad_norm": 25.72220230102539,
"learning_rate": 4.450000000000001e-06,
"loss": 1.0472,
"step": 89
},
{
"epoch": 0.4084720121028744,
"grad_norm": 3.1196579933166504,
"learning_rate": 4.5e-06,
"loss": 1.0055,
"step": 90
},
{
"epoch": 0.4130105900151286,
"grad_norm": 4.584676742553711,
"learning_rate": 4.5500000000000005e-06,
"loss": 0.9848,
"step": 91
},
{
"epoch": 0.4175491679273828,
"grad_norm": 6.140016078948975,
"learning_rate": 4.600000000000001e-06,
"loss": 0.948,
"step": 92
},
{
"epoch": 0.42208774583963693,
"grad_norm": 5.266363620758057,
"learning_rate": 4.65e-06,
"loss": 0.9261,
"step": 93
},
{
"epoch": 0.4266263237518911,
"grad_norm": 3.6768720149993896,
"learning_rate": 4.7e-06,
"loss": 0.9772,
"step": 94
},
{
"epoch": 0.43116490166414523,
"grad_norm": 8.903961181640625,
"learning_rate": 4.75e-06,
"loss": 0.9514,
"step": 95
},
{
"epoch": 0.4357034795763994,
"grad_norm": 15.684305191040039,
"learning_rate": 4.800000000000001e-06,
"loss": 0.9629,
"step": 96
},
{
"epoch": 0.4402420574886536,
"grad_norm": 21.329519271850586,
"learning_rate": 4.85e-06,
"loss": 1.0118,
"step": 97
},
{
"epoch": 0.44478063540090773,
"grad_norm": 7.824005603790283,
"learning_rate": 4.9000000000000005e-06,
"loss": 0.9094,
"step": 98
},
{
"epoch": 0.4493192133131619,
"grad_norm": 2.283174514770508,
"learning_rate": 4.95e-06,
"loss": 0.8877,
"step": 99
},
{
"epoch": 0.45385779122541603,
"grad_norm": 5.028634548187256,
"learning_rate": 5e-06,
"loss": 0.8946,
"step": 100
},
{
"epoch": 0.4583963691376702,
"grad_norm": 6.072951793670654,
"learning_rate": 4.9999917112344245e-06,
"loss": 0.9248,
"step": 101
},
{
"epoch": 0.4629349470499244,
"grad_norm": 7.111969470977783,
"learning_rate": 4.999966844992657e-06,
"loss": 0.9172,
"step": 102
},
{
"epoch": 0.46747352496217853,
"grad_norm": 3.792372465133667,
"learning_rate": 4.999925401439588e-06,
"loss": 0.8718,
"step": 103
},
{
"epoch": 0.4720121028744327,
"grad_norm": 4.017702579498291,
"learning_rate": 4.999867380850031e-06,
"loss": 0.9121,
"step": 104
},
{
"epoch": 0.47655068078668683,
"grad_norm": 6.903019905090332,
"learning_rate": 4.99979278360872e-06,
"loss": 0.9359,
"step": 105
},
{
"epoch": 0.481089258698941,
"grad_norm": 3.740607500076294,
"learning_rate": 4.999701610210309e-06,
"loss": 0.8474,
"step": 106
},
{
"epoch": 0.4856278366111952,
"grad_norm": 3.379190683364868,
"learning_rate": 4.999593861259373e-06,
"loss": 0.8641,
"step": 107
},
{
"epoch": 0.49016641452344933,
"grad_norm": 3.0796492099761963,
"learning_rate": 4.999469537470394e-06,
"loss": 0.8255,
"step": 108
},
{
"epoch": 0.4947049924357035,
"grad_norm": 3.3708953857421875,
"learning_rate": 4.999328639667765e-06,
"loss": 0.8823,
"step": 109
},
{
"epoch": 0.49924357034795763,
"grad_norm": 3.2106549739837646,
"learning_rate": 4.999171168785783e-06,
"loss": 0.8584,
"step": 110
},
{
"epoch": 0.5037821482602118,
"grad_norm": 6.520877838134766,
"learning_rate": 4.998997125868638e-06,
"loss": 0.8718,
"step": 111
},
{
"epoch": 0.5083207261724659,
"grad_norm": 1.0179104804992676,
"learning_rate": 4.99880651207041e-06,
"loss": 0.871,
"step": 112
},
{
"epoch": 0.5128593040847201,
"grad_norm": 1.1883283853530884,
"learning_rate": 4.998599328655063e-06,
"loss": 0.8689,
"step": 113
},
{
"epoch": 0.5173978819969742,
"grad_norm": 1.2722241878509521,
"learning_rate": 4.998375576996431e-06,
"loss": 0.8594,
"step": 114
},
{
"epoch": 0.5219364599092284,
"grad_norm": 1.8931000232696533,
"learning_rate": 4.9981352585782154e-06,
"loss": 0.8327,
"step": 115
},
{
"epoch": 0.5264750378214826,
"grad_norm": 3.093480348587036,
"learning_rate": 4.997878374993971e-06,
"loss": 0.8372,
"step": 116
},
{
"epoch": 0.5310136157337367,
"grad_norm": 5.2650957107543945,
"learning_rate": 4.9976049279470955e-06,
"loss": 0.8386,
"step": 117
},
{
"epoch": 0.5355521936459909,
"grad_norm": 4.467101573944092,
"learning_rate": 4.997314919250818e-06,
"loss": 0.7788,
"step": 118
},
{
"epoch": 0.540090771558245,
"grad_norm": 3.614868640899658,
"learning_rate": 4.997008350828192e-06,
"loss": 0.874,
"step": 119
},
{
"epoch": 0.5446293494704992,
"grad_norm": 12.209319114685059,
"learning_rate": 4.996685224712077e-06,
"loss": 0.8223,
"step": 120
},
{
"epoch": 0.5491679273827534,
"grad_norm": 4.1903252601623535,
"learning_rate": 4.9963455430451245e-06,
"loss": 0.8455,
"step": 121
},
{
"epoch": 0.5537065052950075,
"grad_norm": 2.649549722671509,
"learning_rate": 4.9959893080797675e-06,
"loss": 0.7875,
"step": 122
},
{
"epoch": 0.5582450832072617,
"grad_norm": 4.935393810272217,
"learning_rate": 4.995616522178207e-06,
"loss": 0.8691,
"step": 123
},
{
"epoch": 0.5627836611195158,
"grad_norm": 5.565356731414795,
"learning_rate": 4.995227187812389e-06,
"loss": 0.8267,
"step": 124
},
{
"epoch": 0.56732223903177,
"grad_norm": 5.23801851272583,
"learning_rate": 4.994821307563995e-06,
"loss": 0.826,
"step": 125
},
{
"epoch": 0.5718608169440242,
"grad_norm": 4.2979736328125,
"learning_rate": 4.994398884124422e-06,
"loss": 0.8068,
"step": 126
},
{
"epoch": 0.5763993948562783,
"grad_norm": 6.285053730010986,
"learning_rate": 4.993959920294764e-06,
"loss": 0.819,
"step": 127
},
{
"epoch": 0.5809379727685325,
"grad_norm": 10.453060150146484,
"learning_rate": 4.9935044189857975e-06,
"loss": 0.773,
"step": 128
},
{
"epoch": 0.5854765506807866,
"grad_norm": 31.77552604675293,
"learning_rate": 4.993032383217957e-06,
"loss": 0.8738,
"step": 129
},
{
"epoch": 0.5900151285930408,
"grad_norm": 1.1456351280212402,
"learning_rate": 4.992543816121317e-06,
"loss": 0.7518,
"step": 130
},
{
"epoch": 0.594553706505295,
"grad_norm": 1.3343666791915894,
"learning_rate": 4.992038720935572e-06,
"loss": 0.8108,
"step": 131
},
{
"epoch": 0.5990922844175491,
"grad_norm": 0.8912076354026794,
"learning_rate": 4.991517101010015e-06,
"loss": 0.8159,
"step": 132
},
{
"epoch": 0.6036308623298033,
"grad_norm": 2.249366044998169,
"learning_rate": 4.990978959803513e-06,
"loss": 0.8124,
"step": 133
},
{
"epoch": 0.6081694402420574,
"grad_norm": 1.2822734117507935,
"learning_rate": 4.990424300884488e-06,
"loss": 0.8213,
"step": 134
},
{
"epoch": 0.6127080181543116,
"grad_norm": 6.053490161895752,
"learning_rate": 4.98985312793089e-06,
"loss": 0.7829,
"step": 135
},
{
"epoch": 0.6172465960665658,
"grad_norm": 2.2606236934661865,
"learning_rate": 4.989265444730176e-06,
"loss": 0.7729,
"step": 136
},
{
"epoch": 0.6217851739788199,
"grad_norm": 3.8894989490509033,
"learning_rate": 4.988661255179276e-06,
"loss": 0.774,
"step": 137
},
{
"epoch": 0.6263237518910741,
"grad_norm": 5.648194313049316,
"learning_rate": 4.988040563284582e-06,
"loss": 0.7251,
"step": 138
},
{
"epoch": 0.6308623298033282,
"grad_norm": 1.7583339214324951,
"learning_rate": 4.98740337316191e-06,
"loss": 0.7875,
"step": 139
},
{
"epoch": 0.6354009077155824,
"grad_norm": 1.3959944248199463,
"learning_rate": 4.9867496890364734e-06,
"loss": 0.7726,
"step": 140
},
{
"epoch": 0.6399394856278366,
"grad_norm": 1.7879443168640137,
"learning_rate": 4.986079515242861e-06,
"loss": 0.786,
"step": 141
},
{
"epoch": 0.6444780635400907,
"grad_norm": 2.0919816493988037,
"learning_rate": 4.985392856225003e-06,
"loss": 0.7802,
"step": 142
},
{
"epoch": 0.649016641452345,
"grad_norm": 1.272477626800537,
"learning_rate": 4.984689716536145e-06,
"loss": 0.7842,
"step": 143
},
{
"epoch": 0.653555219364599,
"grad_norm": 1.1265331506729126,
"learning_rate": 4.983970100838814e-06,
"loss": 0.736,
"step": 144
},
{
"epoch": 0.6580937972768532,
"grad_norm": 0.8514362573623657,
"learning_rate": 4.983234013904791e-06,
"loss": 0.749,
"step": 145
},
{
"epoch": 0.6626323751891074,
"grad_norm": 3.148453950881958,
"learning_rate": 4.9824814606150774e-06,
"loss": 0.7884,
"step": 146
},
{
"epoch": 0.6671709531013615,
"grad_norm": 1.531554937362671,
"learning_rate": 4.981712445959864e-06,
"loss": 0.767,
"step": 147
},
{
"epoch": 0.6717095310136157,
"grad_norm": 1.7979401350021362,
"learning_rate": 4.980926975038496e-06,
"loss": 0.7575,
"step": 148
},
{
"epoch": 0.6762481089258698,
"grad_norm": 1.1131621599197388,
"learning_rate": 4.9801250530594415e-06,
"loss": 0.76,
"step": 149
},
{
"epoch": 0.680786686838124,
"grad_norm": 1.2112400531768799,
"learning_rate": 4.9793066853402535e-06,
"loss": 0.769,
"step": 150
},
{
"epoch": 0.6853252647503782,
"grad_norm": 1.276172161102295,
"learning_rate": 4.978471877307541e-06,
"loss": 0.7641,
"step": 151
},
{
"epoch": 0.6898638426626323,
"grad_norm": 7.614717960357666,
"learning_rate": 4.977620634496926e-06,
"loss": 0.7614,
"step": 152
},
{
"epoch": 0.6944024205748865,
"grad_norm": 0.9541272521018982,
"learning_rate": 4.976752962553008e-06,
"loss": 0.7406,
"step": 153
},
{
"epoch": 0.6989409984871406,
"grad_norm": 0.9793027639389038,
"learning_rate": 4.975868867229332e-06,
"loss": 0.7538,
"step": 154
},
{
"epoch": 0.7034795763993948,
"grad_norm": 8.540267944335938,
"learning_rate": 4.974968354388346e-06,
"loss": 0.7616,
"step": 155
},
{
"epoch": 0.708018154311649,
"grad_norm": 2.279240608215332,
"learning_rate": 4.97405143000136e-06,
"loss": 0.7157,
"step": 156
},
{
"epoch": 0.7125567322239031,
"grad_norm": 2.947227716445923,
"learning_rate": 4.973118100148513e-06,
"loss": 0.7348,
"step": 157
},
{
"epoch": 0.7170953101361573,
"grad_norm": 2.211785316467285,
"learning_rate": 4.9721683710187255e-06,
"loss": 0.7144,
"step": 158
},
{
"epoch": 0.7216338880484114,
"grad_norm": 1.3755372762680054,
"learning_rate": 4.971202248909662e-06,
"loss": 0.6857,
"step": 159
},
{
"epoch": 0.7261724659606656,
"grad_norm": 2.564708709716797,
"learning_rate": 4.970219740227693e-06,
"loss": 0.7124,
"step": 160
},
{
"epoch": 0.7307110438729199,
"grad_norm": 11.76566219329834,
"learning_rate": 4.9692208514878445e-06,
"loss": 0.7815,
"step": 161
},
{
"epoch": 0.735249621785174,
"grad_norm": 3.7665200233459473,
"learning_rate": 4.9682055893137605e-06,
"loss": 0.7021,
"step": 162
},
{
"epoch": 0.7397881996974282,
"grad_norm": 2.9076344966888428,
"learning_rate": 4.967173960437657e-06,
"loss": 0.7083,
"step": 163
},
{
"epoch": 0.7443267776096822,
"grad_norm": 2.336026191711426,
"learning_rate": 4.966125971700277e-06,
"loss": 0.7455,
"step": 164
},
{
"epoch": 0.7488653555219364,
"grad_norm": 6.48813009262085,
"learning_rate": 4.965061630050848e-06,
"loss": 0.7628,
"step": 165
},
{
"epoch": 0.7534039334341907,
"grad_norm": 1.9123872518539429,
"learning_rate": 4.9639809425470324e-06,
"loss": 0.7212,
"step": 166
},
{
"epoch": 0.7579425113464447,
"grad_norm": 6.638424873352051,
"learning_rate": 4.962883916354882e-06,
"loss": 0.7523,
"step": 167
},
{
"epoch": 0.762481089258699,
"grad_norm": 1.3606716394424438,
"learning_rate": 4.961770558748793e-06,
"loss": 0.7651,
"step": 168
},
{
"epoch": 0.7670196671709532,
"grad_norm": 1.2726523876190186,
"learning_rate": 4.960640877111451e-06,
"loss": 0.7729,
"step": 169
},
{
"epoch": 0.7715582450832073,
"grad_norm": 0.9216910600662231,
"learning_rate": 4.959494878933792e-06,
"loss": 0.7162,
"step": 170
},
{
"epoch": 0.7760968229954615,
"grad_norm": 2.3702430725097656,
"learning_rate": 4.958332571814941e-06,
"loss": 0.7038,
"step": 171
},
{
"epoch": 0.7806354009077155,
"grad_norm": 4.6101555824279785,
"learning_rate": 4.957153963462172e-06,
"loss": 0.7141,
"step": 172
},
{
"epoch": 0.7851739788199698,
"grad_norm": 1.7314170598983765,
"learning_rate": 4.955959061690853e-06,
"loss": 0.7374,
"step": 173
},
{
"epoch": 0.789712556732224,
"grad_norm": 2.91019868850708,
"learning_rate": 4.9547478744243914e-06,
"loss": 0.7048,
"step": 174
},
{
"epoch": 0.794251134644478,
"grad_norm": 2.872775077819824,
"learning_rate": 4.953520409694186e-06,
"loss": 0.6912,
"step": 175
},
{
"epoch": 0.7987897125567323,
"grad_norm": 2.2696948051452637,
"learning_rate": 4.952276675639569e-06,
"loss": 0.7432,
"step": 176
},
{
"epoch": 0.8033282904689864,
"grad_norm": 1.8855810165405273,
"learning_rate": 4.951016680507757e-06,
"loss": 0.7056,
"step": 177
},
{
"epoch": 0.8078668683812406,
"grad_norm": 5.202772617340088,
"learning_rate": 4.9497404326537954e-06,
"loss": 0.7114,
"step": 178
},
{
"epoch": 0.8124054462934948,
"grad_norm": 0.7916449904441833,
"learning_rate": 4.948447940540497e-06,
"loss": 0.7202,
"step": 179
},
{
"epoch": 0.8169440242057489,
"grad_norm": 1.5009609460830688,
"learning_rate": 4.947139212738395e-06,
"loss": 0.7245,
"step": 180
},
{
"epoch": 0.8214826021180031,
"grad_norm": 1.858067512512207,
"learning_rate": 4.945814257925679e-06,
"loss": 0.6962,
"step": 181
},
{
"epoch": 0.8260211800302572,
"grad_norm": 0.7835391163825989,
"learning_rate": 4.94447308488814e-06,
"loss": 0.6875,
"step": 182
},
{
"epoch": 0.8305597579425114,
"grad_norm": 21.755929946899414,
"learning_rate": 4.943115702519115e-06,
"loss": 0.7304,
"step": 183
},
{
"epoch": 0.8350983358547656,
"grad_norm": 6.944667816162109,
"learning_rate": 4.941742119819421e-06,
"loss": 0.7381,
"step": 184
},
{
"epoch": 0.8396369137670197,
"grad_norm": 0.9813210368156433,
"learning_rate": 4.940352345897304e-06,
"loss": 0.682,
"step": 185
},
{
"epoch": 0.8441754916792739,
"grad_norm": 1.8010449409484863,
"learning_rate": 4.938946389968372e-06,
"loss": 0.7639,
"step": 186
},
{
"epoch": 0.848714069591528,
"grad_norm": 2.293980121612549,
"learning_rate": 4.937524261355535e-06,
"loss": 0.7027,
"step": 187
},
{
"epoch": 0.8532526475037822,
"grad_norm": 1.9937771558761597,
"learning_rate": 4.9360859694889475e-06,
"loss": 0.688,
"step": 188
},
{
"epoch": 0.8577912254160364,
"grad_norm": 1.8442484140396118,
"learning_rate": 4.934631523905938e-06,
"loss": 0.6261,
"step": 189
},
{
"epoch": 0.8623298033282905,
"grad_norm": 9.626107215881348,
"learning_rate": 4.933160934250957e-06,
"loss": 0.7605,
"step": 190
},
{
"epoch": 0.8668683812405447,
"grad_norm": 14.051777839660645,
"learning_rate": 4.931674210275499e-06,
"loss": 0.7183,
"step": 191
},
{
"epoch": 0.8714069591527988,
"grad_norm": 2.1191225051879883,
"learning_rate": 4.930171361838052e-06,
"loss": 0.697,
"step": 192
},
{
"epoch": 0.875945537065053,
"grad_norm": 1.4051584005355835,
"learning_rate": 4.928652398904022e-06,
"loss": 0.6985,
"step": 193
},
{
"epoch": 0.8804841149773072,
"grad_norm": 0.7633137106895447,
"learning_rate": 4.92711733154567e-06,
"loss": 0.7018,
"step": 194
},
{
"epoch": 0.8850226928895613,
"grad_norm": 1.5257729291915894,
"learning_rate": 4.925566169942048e-06,
"loss": 0.7108,
"step": 195
},
{
"epoch": 0.8895612708018155,
"grad_norm": 1.7885994911193848,
"learning_rate": 4.9239989243789275e-06,
"loss": 0.7251,
"step": 196
},
{
"epoch": 0.8940998487140696,
"grad_norm": 1.8133364915847778,
"learning_rate": 4.922415605248734e-06,
"loss": 0.691,
"step": 197
},
{
"epoch": 0.8986384266263238,
"grad_norm": 1.3306565284729004,
"learning_rate": 4.920816223050475e-06,
"loss": 0.6496,
"step": 198
},
{
"epoch": 0.903177004538578,
"grad_norm": 2.5053746700286865,
"learning_rate": 4.919200788389675e-06,
"loss": 0.7174,
"step": 199
},
{
"epoch": 0.9077155824508321,
"grad_norm": 2.8250479698181152,
"learning_rate": 4.917569311978301e-06,
"loss": 0.6992,
"step": 200
},
{
"epoch": 0.9122541603630863,
"grad_norm": 0.6993988752365112,
"learning_rate": 4.915921804634693e-06,
"loss": 0.6983,
"step": 201
},
{
"epoch": 0.9167927382753404,
"grad_norm": 2.591536045074463,
"learning_rate": 4.914258277283494e-06,
"loss": 0.6686,
"step": 202
},
{
"epoch": 0.9213313161875946,
"grad_norm": 1.7433080673217773,
"learning_rate": 4.912578740955573e-06,
"loss": 0.686,
"step": 203
},
{
"epoch": 0.9258698940998488,
"grad_norm": 1.5048431158065796,
"learning_rate": 4.910883206787958e-06,
"loss": 0.7043,
"step": 204
},
{
"epoch": 0.9304084720121029,
"grad_norm": 0.7160290479660034,
"learning_rate": 4.9091716860237545e-06,
"loss": 0.6703,
"step": 205
},
{
"epoch": 0.9349470499243571,
"grad_norm": 16.788084030151367,
"learning_rate": 4.907444190012081e-06,
"loss": 0.7465,
"step": 206
},
{
"epoch": 0.9394856278366112,
"grad_norm": 6.705326557159424,
"learning_rate": 4.905700730207983e-06,
"loss": 0.6692,
"step": 207
},
{
"epoch": 0.9440242057488654,
"grad_norm": 1.1539785861968994,
"learning_rate": 4.903941318172365e-06,
"loss": 0.6769,
"step": 208
},
{
"epoch": 0.9485627836611196,
"grad_norm": 0.806441068649292,
"learning_rate": 4.902165965571911e-06,
"loss": 0.6788,
"step": 209
},
{
"epoch": 0.9531013615733737,
"grad_norm": 2.599201202392578,
"learning_rate": 4.900374684179005e-06,
"loss": 0.6845,
"step": 210
},
{
"epoch": 0.9576399394856279,
"grad_norm": 1.9495007991790771,
"learning_rate": 4.898567485871656e-06,
"loss": 0.68,
"step": 211
},
{
"epoch": 0.962178517397882,
"grad_norm": 1.3471159934997559,
"learning_rate": 4.896744382633419e-06,
"loss": 0.6799,
"step": 212
},
{
"epoch": 0.9667170953101362,
"grad_norm": 2.436737537384033,
"learning_rate": 4.894905386553316e-06,
"loss": 0.6688,
"step": 213
},
{
"epoch": 0.9712556732223904,
"grad_norm": 2.486992597579956,
"learning_rate": 4.893050509825749e-06,
"loss": 0.6866,
"step": 214
},
{
"epoch": 0.9757942511346445,
"grad_norm": 2.137023687362671,
"learning_rate": 4.891179764750434e-06,
"loss": 0.6891,
"step": 215
},
{
"epoch": 0.9803328290468987,
"grad_norm": 1.3739172220230103,
"learning_rate": 4.8892931637323e-06,
"loss": 0.6219,
"step": 216
},
{
"epoch": 0.9848714069591528,
"grad_norm": 1.6083399057388306,
"learning_rate": 4.887390719281423e-06,
"loss": 0.6805,
"step": 217
},
{
"epoch": 0.989409984871407,
"grad_norm": 0.7767590284347534,
"learning_rate": 4.885472444012937e-06,
"loss": 0.6857,
"step": 218
},
{
"epoch": 0.9939485627836612,
"grad_norm": 0.8085631728172302,
"learning_rate": 4.883538350646949e-06,
"loss": 0.6594,
"step": 219
},
{
"epoch": 0.9984871406959153,
"grad_norm": 0.7762560844421387,
"learning_rate": 4.881588452008457e-06,
"loss": 0.6972,
"step": 220
},
{
"epoch": 1.0,
"grad_norm": 0.7762560844421387,
"learning_rate": 4.8796227610272615e-06,
"loss": 0.2152,
"step": 221
},
{
"epoch": 1.0045385779122542,
"grad_norm": 0.8768157958984375,
"learning_rate": 4.8776412907378845e-06,
"loss": 0.6847,
"step": 222
},
{
"epoch": 1.0090771558245084,
"grad_norm": 2.3720204830169678,
"learning_rate": 4.8756440542794805e-06,
"loss": 0.6646,
"step": 223
},
{
"epoch": 1.0136157337367624,
"grad_norm": 2.1314537525177,
"learning_rate": 4.873631064895749e-06,
"loss": 0.676,
"step": 224
},
{
"epoch": 1.0181543116490166,
"grad_norm": 1.7491836547851562,
"learning_rate": 4.871602335934847e-06,
"loss": 0.6474,
"step": 225
},
{
"epoch": 1.0226928895612708,
"grad_norm": 1.7630263566970825,
"learning_rate": 4.8695578808493034e-06,
"loss": 0.6541,
"step": 226
},
{
"epoch": 1.027231467473525,
"grad_norm": 1.195518970489502,
"learning_rate": 4.867497713195925e-06,
"loss": 0.6529,
"step": 227
},
{
"epoch": 1.0317700453857792,
"grad_norm": 5.6356377601623535,
"learning_rate": 4.8654218466357066e-06,
"loss": 0.6666,
"step": 228
},
{
"epoch": 1.0363086232980332,
"grad_norm": 21.327880859375,
"learning_rate": 4.863330294933748e-06,
"loss": 0.6721,
"step": 229
},
{
"epoch": 1.0408472012102874,
"grad_norm": 14.34103012084961,
"learning_rate": 4.8612230719591535e-06,
"loss": 0.6374,
"step": 230
},
{
"epoch": 1.0453857791225416,
"grad_norm": 7.112085819244385,
"learning_rate": 4.859100191684946e-06,
"loss": 0.6729,
"step": 231
},
{
"epoch": 1.0499243570347958,
"grad_norm": 2.3837637901306152,
"learning_rate": 4.856961668187968e-06,
"loss": 0.6741,
"step": 232
},
{
"epoch": 1.05446293494705,
"grad_norm": 1.5125519037246704,
"learning_rate": 4.854807515648799e-06,
"loss": 0.6584,
"step": 233
},
{
"epoch": 1.059001512859304,
"grad_norm": 1.8248745203018188,
"learning_rate": 4.852637748351651e-06,
"loss": 0.6481,
"step": 234
},
{
"epoch": 1.0635400907715582,
"grad_norm": 2.260824203491211,
"learning_rate": 4.850452380684275e-06,
"loss": 0.6695,
"step": 235
},
{
"epoch": 1.0680786686838124,
"grad_norm": 1.764312982559204,
"learning_rate": 4.848251427137875e-06,
"loss": 0.6638,
"step": 236
},
{
"epoch": 1.0726172465960666,
"grad_norm": 0.9826205968856812,
"learning_rate": 4.846034902306997e-06,
"loss": 0.6515,
"step": 237
},
{
"epoch": 1.0771558245083208,
"grad_norm": 1.1290264129638672,
"learning_rate": 4.8438028208894496e-06,
"loss": 0.6483,
"step": 238
},
{
"epoch": 1.0816944024205748,
"grad_norm": 4.97009801864624,
"learning_rate": 4.841555197686189e-06,
"loss": 0.6605,
"step": 239
},
{
"epoch": 1.086232980332829,
"grad_norm": 6.06207799911499,
"learning_rate": 4.839292047601234e-06,
"loss": 0.6147,
"step": 240
},
{
"epoch": 1.0907715582450832,
"grad_norm": 1.2093101739883423,
"learning_rate": 4.837013385641562e-06,
"loss": 0.6739,
"step": 241
},
{
"epoch": 1.0953101361573374,
"grad_norm": 0.9535529017448425,
"learning_rate": 4.834719226917007e-06,
"loss": 0.6392,
"step": 242
},
{
"epoch": 1.0998487140695916,
"grad_norm": 1.4470645189285278,
"learning_rate": 4.832409586640164e-06,
"loss": 0.6357,
"step": 243
},
{
"epoch": 1.1043872919818456,
"grad_norm": 1.4899321794509888,
"learning_rate": 4.830084480126288e-06,
"loss": 0.6704,
"step": 244
},
{
"epoch": 1.1089258698940998,
"grad_norm": 0.7240656614303589,
"learning_rate": 4.827743922793189e-06,
"loss": 0.621,
"step": 245
},
{
"epoch": 1.113464447806354,
"grad_norm": 0.8096688985824585,
"learning_rate": 4.8253879301611315e-06,
"loss": 0.6561,
"step": 246
},
{
"epoch": 1.1180030257186082,
"grad_norm": 1.7580249309539795,
"learning_rate": 4.823016517852731e-06,
"loss": 0.5893,
"step": 247
},
{
"epoch": 1.1225416036308624,
"grad_norm": 2.5382940769195557,
"learning_rate": 4.820629701592853e-06,
"loss": 0.6548,
"step": 248
},
{
"epoch": 1.1270801815431164,
"grad_norm": 1.0767178535461426,
"learning_rate": 4.8182274972085065e-06,
"loss": 0.6801,
"step": 249
},
{
"epoch": 1.1316187594553706,
"grad_norm": 0.7919514179229736,
"learning_rate": 4.815809920628738e-06,
"loss": 0.6314,
"step": 250
},
{
"epoch": 1.1361573373676248,
"grad_norm": 1.305253267288208,
"learning_rate": 4.813376987884527e-06,
"loss": 0.6347,
"step": 251
},
{
"epoch": 1.140695915279879,
"grad_norm": 1.4656856060028076,
"learning_rate": 4.810928715108683e-06,
"loss": 0.6253,
"step": 252
},
{
"epoch": 1.1452344931921332,
"grad_norm": 1.2813221216201782,
"learning_rate": 4.808465118535732e-06,
"loss": 0.6751,
"step": 253
},
{
"epoch": 1.1497730711043872,
"grad_norm": 3.507342576980591,
"learning_rate": 4.805986214501813e-06,
"loss": 0.6606,
"step": 254
},
{
"epoch": 1.1543116490166414,
"grad_norm": 4.23391056060791,
"learning_rate": 4.803492019444571e-06,
"loss": 0.6278,
"step": 255
},
{
"epoch": 1.1588502269288956,
"grad_norm": 2.3074967861175537,
"learning_rate": 4.8009825499030426e-06,
"loss": 0.6175,
"step": 256
},
{
"epoch": 1.1633888048411498,
"grad_norm": 1.3244863748550415,
"learning_rate": 4.798457822517554e-06,
"loss": 0.6392,
"step": 257
},
{
"epoch": 1.167927382753404,
"grad_norm": 1.0530226230621338,
"learning_rate": 4.795917854029601e-06,
"loss": 0.6305,
"step": 258
},
{
"epoch": 1.172465960665658,
"grad_norm": 2.187415599822998,
"learning_rate": 4.79336266128175e-06,
"loss": 0.6432,
"step": 259
},
{
"epoch": 1.1770045385779122,
"grad_norm": 1.4672960042953491,
"learning_rate": 4.790792261217513e-06,
"loss": 0.649,
"step": 260
},
{
"epoch": 1.1815431164901664,
"grad_norm": 1.1620965003967285,
"learning_rate": 4.788206670881245e-06,
"loss": 0.6507,
"step": 261
},
{
"epoch": 1.1860816944024206,
"grad_norm": 0.7209274768829346,
"learning_rate": 4.785605907418029e-06,
"loss": 0.6502,
"step": 262
},
{
"epoch": 1.1906202723146748,
"grad_norm": 3.6349446773529053,
"learning_rate": 4.78298998807356e-06,
"loss": 0.6251,
"step": 263
},
{
"epoch": 1.1951588502269288,
"grad_norm": 13.263801574707031,
"learning_rate": 4.7803589301940306e-06,
"loss": 0.6663,
"step": 264
},
{
"epoch": 1.199697428139183,
"grad_norm": 12.560731887817383,
"learning_rate": 4.777712751226019e-06,
"loss": 0.6709,
"step": 265
},
{
"epoch": 1.2042360060514372,
"grad_norm": 3.721285820007324,
"learning_rate": 4.775051468716371e-06,
"loss": 0.6555,
"step": 266
},
{
"epoch": 1.2087745839636914,
"grad_norm": 0.974590539932251,
"learning_rate": 4.772375100312084e-06,
"loss": 0.6308,
"step": 267
},
{
"epoch": 1.2133131618759456,
"grad_norm": 1.3410248756408691,
"learning_rate": 4.769683663760191e-06,
"loss": 0.6506,
"step": 268
},
{
"epoch": 1.2178517397881996,
"grad_norm": 1.4115511178970337,
"learning_rate": 4.7669771769076395e-06,
"loss": 0.6296,
"step": 269
},
{
"epoch": 1.2223903177004538,
"grad_norm": 1.355098843574524,
"learning_rate": 4.764255657701179e-06,
"loss": 0.6756,
"step": 270
},
{
"epoch": 1.226928895612708,
"grad_norm": 0.9310200810432434,
"learning_rate": 4.761519124187237e-06,
"loss": 0.6724,
"step": 271
},
{
"epoch": 1.2314674735249622,
"grad_norm": 1.0098122358322144,
"learning_rate": 4.758767594511801e-06,
"loss": 0.6595,
"step": 272
},
{
"epoch": 1.2360060514372164,
"grad_norm": 2.7444238662719727,
"learning_rate": 4.7560010869202985e-06,
"loss": 0.582,
"step": 273
},
{
"epoch": 1.2405446293494704,
"grad_norm": 2.2694830894470215,
"learning_rate": 4.753219619757477e-06,
"loss": 0.6411,
"step": 274
},
{
"epoch": 1.2450832072617246,
"grad_norm": 1.1762354373931885,
"learning_rate": 4.750423211467278e-06,
"loss": 0.6358,
"step": 275
},
{
"epoch": 1.2496217851739788,
"grad_norm": 0.86478191614151,
"learning_rate": 4.7476118805927214e-06,
"loss": 0.6234,
"step": 276
},
{
"epoch": 1.254160363086233,
"grad_norm": 1.143272876739502,
"learning_rate": 4.7447856457757765e-06,
"loss": 0.6627,
"step": 277
},
{
"epoch": 1.258698940998487,
"grad_norm": 1.7226762771606445,
"learning_rate": 4.7419445257572414e-06,
"loss": 0.6248,
"step": 278
},
{
"epoch": 1.2632375189107412,
"grad_norm": 1.428463101387024,
"learning_rate": 4.739088539376618e-06,
"loss": 0.6577,
"step": 279
},
{
"epoch": 1.2677760968229954,
"grad_norm": 0.9464501142501831,
"learning_rate": 4.736217705571989e-06,
"loss": 0.6464,
"step": 280
},
{
"epoch": 1.2723146747352496,
"grad_norm": 0.8889546394348145,
"learning_rate": 4.733332043379889e-06,
"loss": 0.6249,
"step": 281
},
{
"epoch": 1.2768532526475038,
"grad_norm": 0.7456269860267639,
"learning_rate": 4.730431571935178e-06,
"loss": 0.6242,
"step": 282
},
{
"epoch": 1.281391830559758,
"grad_norm": 9.802299499511719,
"learning_rate": 4.72751631047092e-06,
"loss": 0.6576,
"step": 283
},
{
"epoch": 1.2859304084720122,
"grad_norm": 15.863835334777832,
"learning_rate": 4.72458627831825e-06,
"loss": 0.6916,
"step": 284
},
{
"epoch": 1.2904689863842662,
"grad_norm": 15.025418281555176,
"learning_rate": 4.721641494906247e-06,
"loss": 0.7036,
"step": 285
},
{
"epoch": 1.2950075642965204,
"grad_norm": 3.8970537185668945,
"learning_rate": 4.718681979761806e-06,
"loss": 0.6166,
"step": 286
},
{
"epoch": 1.2995461422087746,
"grad_norm": 0.6507979035377502,
"learning_rate": 4.715707752509512e-06,
"loss": 0.613,
"step": 287
},
{
"epoch": 1.3040847201210286,
"grad_norm": 1.1878042221069336,
"learning_rate": 4.712718832871499e-06,
"loss": 0.6474,
"step": 288
},
{
"epoch": 1.3086232980332828,
"grad_norm": 1.0940614938735962,
"learning_rate": 4.709715240667332e-06,
"loss": 0.6577,
"step": 289
},
{
"epoch": 1.313161875945537,
"grad_norm": 0.9987061619758606,
"learning_rate": 4.706696995813869e-06,
"loss": 0.6571,
"step": 290
},
{
"epoch": 1.3177004538577912,
"grad_norm": 1.5589380264282227,
"learning_rate": 4.7036641183251285e-06,
"loss": 0.6495,
"step": 291
},
{
"epoch": 1.3222390317700454,
"grad_norm": 1.525474190711975,
"learning_rate": 4.700616628312159e-06,
"loss": 0.5986,
"step": 292
},
{
"epoch": 1.3267776096822996,
"grad_norm": 0.8548336625099182,
"learning_rate": 4.697554545982904e-06,
"loss": 0.6034,
"step": 293
},
{
"epoch": 1.3313161875945538,
"grad_norm": 4.231250286102295,
"learning_rate": 4.6944778916420705e-06,
"loss": 0.6405,
"step": 294
},
{
"epoch": 1.3358547655068078,
"grad_norm": 8.273162841796875,
"learning_rate": 4.691386685690993e-06,
"loss": 0.6635,
"step": 295
},
{
"epoch": 1.340393343419062,
"grad_norm": 4.974193096160889,
"learning_rate": 4.6882809486274934e-06,
"loss": 0.6289,
"step": 296
},
{
"epoch": 1.3449319213313162,
"grad_norm": 3.757338523864746,
"learning_rate": 4.685160701045757e-06,
"loss": 0.6227,
"step": 297
},
{
"epoch": 1.3494704992435702,
"grad_norm": 1.2015799283981323,
"learning_rate": 4.68202596363618e-06,
"loss": 0.6237,
"step": 298
},
{
"epoch": 1.3540090771558244,
"grad_norm": 0.7638722658157349,
"learning_rate": 4.678876757185248e-06,
"loss": 0.6063,
"step": 299
},
{
"epoch": 1.3585476550680786,
"grad_norm": 1.2864232063293457,
"learning_rate": 4.675713102575389e-06,
"loss": 0.5997,
"step": 300
},
{
"epoch": 1.3630862329803328,
"grad_norm": 1.1902930736541748,
"learning_rate": 4.672535020784833e-06,
"loss": 0.6352,
"step": 301
},
{
"epoch": 1.367624810892587,
"grad_norm": 1.4321516752243042,
"learning_rate": 4.669342532887482e-06,
"loss": 0.6531,
"step": 302
},
{
"epoch": 1.3721633888048412,
"grad_norm": 1.139543890953064,
"learning_rate": 4.666135660052764e-06,
"loss": 0.6235,
"step": 303
},
{
"epoch": 1.3767019667170954,
"grad_norm": 0.7234447598457336,
"learning_rate": 4.66291442354549e-06,
"loss": 0.6012,
"step": 304
},
{
"epoch": 1.3812405446293494,
"grad_norm": 3.093146324157715,
"learning_rate": 4.659678844725722e-06,
"loss": 0.6058,
"step": 305
},
{
"epoch": 1.3857791225416036,
"grad_norm": 2.840275764465332,
"learning_rate": 4.656428945048622e-06,
"loss": 0.6139,
"step": 306
},
{
"epoch": 1.3903177004538578,
"grad_norm": 1.0061054229736328,
"learning_rate": 4.653164746064315e-06,
"loss": 0.6288,
"step": 307
},
{
"epoch": 1.394856278366112,
"grad_norm": 0.9403374195098877,
"learning_rate": 4.649886269417746e-06,
"loss": 0.6435,
"step": 308
},
{
"epoch": 1.399394856278366,
"grad_norm": 1.0838265419006348,
"learning_rate": 4.646593536848535e-06,
"loss": 0.6485,
"step": 309
},
{
"epoch": 1.4039334341906202,
"grad_norm": 1.2738953828811646,
"learning_rate": 4.643286570190832e-06,
"loss": 0.5993,
"step": 310
},
{
"epoch": 1.4084720121028744,
"grad_norm": 1.3124756813049316,
"learning_rate": 4.639965391373173e-06,
"loss": 0.6154,
"step": 311
},
{
"epoch": 1.4130105900151286,
"grad_norm": 0.7026720643043518,
"learning_rate": 4.636630022418337e-06,
"loss": 0.6493,
"step": 312
},
{
"epoch": 1.4175491679273828,
"grad_norm": 1.101508617401123,
"learning_rate": 4.6332804854431986e-06,
"loss": 0.6437,
"step": 313
},
{
"epoch": 1.422087745839637,
"grad_norm": 0.6824156641960144,
"learning_rate": 4.6299168026585775e-06,
"loss": 0.6017,
"step": 314
},
{
"epoch": 1.426626323751891,
"grad_norm": 0.8083431124687195,
"learning_rate": 4.626538996369096e-06,
"loss": 0.6338,
"step": 315
},
{
"epoch": 1.4311649016641452,
"grad_norm": 0.9624136090278625,
"learning_rate": 4.623147088973031e-06,
"loss": 0.5804,
"step": 316
},
{
"epoch": 1.4357034795763994,
"grad_norm": 0.8000622987747192,
"learning_rate": 4.619741102962161e-06,
"loss": 0.6242,
"step": 317
},
{
"epoch": 1.4402420574886536,
"grad_norm": 1.2038214206695557,
"learning_rate": 4.6163210609216234e-06,
"loss": 0.6259,
"step": 318
},
{
"epoch": 1.4447806354009076,
"grad_norm": 0.8374214768409729,
"learning_rate": 4.612886985529759e-06,
"loss": 0.6078,
"step": 319
},
{
"epoch": 1.4493192133131618,
"grad_norm": 1.0167770385742188,
"learning_rate": 4.609438899557964e-06,
"loss": 0.5972,
"step": 320
},
{
"epoch": 1.453857791225416,
"grad_norm": 0.8266498446464539,
"learning_rate": 4.60597682587054e-06,
"loss": 0.6211,
"step": 321
},
{
"epoch": 1.4583963691376702,
"grad_norm": 0.7585692405700684,
"learning_rate": 4.6025007874245405e-06,
"loss": 0.6233,
"step": 322
},
{
"epoch": 1.4629349470499244,
"grad_norm": 1.6358634233474731,
"learning_rate": 4.59901080726962e-06,
"loss": 0.6075,
"step": 323
},
{
"epoch": 1.4674735249621786,
"grad_norm": 1.1722335815429688,
"learning_rate": 4.595506908547881e-06,
"loss": 0.6066,
"step": 324
},
{
"epoch": 1.4720121028744326,
"grad_norm": 0.9726622104644775,
"learning_rate": 4.591989114493718e-06,
"loss": 0.6506,
"step": 325
},
{
"epoch": 1.4765506807866868,
"grad_norm": 0.8073020577430725,
"learning_rate": 4.588457448433667e-06,
"loss": 0.6077,
"step": 326
},
{
"epoch": 1.481089258698941,
"grad_norm": 0.71394282579422,
"learning_rate": 4.584911933786252e-06,
"loss": 0.5882,
"step": 327
},
{
"epoch": 1.4856278366111952,
"grad_norm": 4.143211364746094,
"learning_rate": 4.581352594061824e-06,
"loss": 0.6047,
"step": 328
},
{
"epoch": 1.4901664145234492,
"grad_norm": 3.5801639556884766,
"learning_rate": 4.5777794528624075e-06,
"loss": 0.6094,
"step": 329
},
{
"epoch": 1.4947049924357034,
"grad_norm": 0.9617034792900085,
"learning_rate": 4.574192533881547e-06,
"loss": 0.6291,
"step": 330
},
{
"epoch": 1.4992435703479576,
"grad_norm": 0.8535535931587219,
"learning_rate": 4.570591860904149e-06,
"loss": 0.6587,
"step": 331
},
{
"epoch": 1.5037821482602118,
"grad_norm": 1.426477074623108,
"learning_rate": 4.566977457806317e-06,
"loss": 0.6347,
"step": 332
},
{
"epoch": 1.508320726172466,
"grad_norm": 1.6053332090377808,
"learning_rate": 4.563349348555207e-06,
"loss": 0.603,
"step": 333
},
{
"epoch": 1.5128593040847202,
"grad_norm": 1.3673542737960815,
"learning_rate": 4.5597075572088545e-06,
"loss": 0.6443,
"step": 334
},
{
"epoch": 1.5173978819969742,
"grad_norm": 1.0444583892822266,
"learning_rate": 4.556052107916023e-06,
"loss": 0.6033,
"step": 335
},
{
"epoch": 1.5219364599092284,
"grad_norm": 2.568854331970215,
"learning_rate": 4.552383024916044e-06,
"loss": 0.6364,
"step": 336
},
{
"epoch": 1.5264750378214826,
"grad_norm": 0.8063260316848755,
"learning_rate": 4.54870033253865e-06,
"loss": 0.6406,
"step": 337
},
{
"epoch": 1.5310136157337366,
"grad_norm": 0.8449574112892151,
"learning_rate": 4.545004055203823e-06,
"loss": 0.5977,
"step": 338
},
{
"epoch": 1.5355521936459908,
"grad_norm": 0.7573151588439941,
"learning_rate": 4.541294217421622e-06,
"loss": 0.6098,
"step": 339
},
{
"epoch": 1.540090771558245,
"grad_norm": 0.7103497982025146,
"learning_rate": 4.537570843792028e-06,
"loss": 0.6344,
"step": 340
},
{
"epoch": 1.5446293494704992,
"grad_norm": 0.7327162623405457,
"learning_rate": 4.5338339590047795e-06,
"loss": 0.6318,
"step": 341
},
{
"epoch": 1.5491679273827534,
"grad_norm": 0.7688593864440918,
"learning_rate": 4.530083587839204e-06,
"loss": 0.6089,
"step": 342
},
{
"epoch": 1.5537065052950076,
"grad_norm": 0.9933049082756042,
"learning_rate": 4.52631975516406e-06,
"loss": 0.6003,
"step": 343
},
{
"epoch": 1.5582450832072618,
"grad_norm": 0.7319652438163757,
"learning_rate": 4.522542485937369e-06,
"loss": 0.6054,
"step": 344
},
{
"epoch": 1.5627836611195158,
"grad_norm": 0.8350914120674133,
"learning_rate": 4.518751805206251e-06,
"loss": 0.606,
"step": 345
},
{
"epoch": 1.56732223903177,
"grad_norm": 1.7987092733383179,
"learning_rate": 4.514947738106755e-06,
"loss": 0.6637,
"step": 346
},
{
"epoch": 1.5718608169440242,
"grad_norm": 0.6338518261909485,
"learning_rate": 4.5111303098637005e-06,
"loss": 0.5778,
"step": 347
},
{
"epoch": 1.5763993948562782,
"grad_norm": 0.854932427406311,
"learning_rate": 4.5072995457905e-06,
"loss": 0.598,
"step": 348
},
{
"epoch": 1.5809379727685324,
"grad_norm": 1.217940330505371,
"learning_rate": 4.503455471288998e-06,
"loss": 0.6087,
"step": 349
},
{
"epoch": 1.5854765506807866,
"grad_norm": 1.329987645149231,
"learning_rate": 4.499598111849299e-06,
"loss": 0.6321,
"step": 350
},
{
"epoch": 1.5900151285930408,
"grad_norm": 1.9689991474151611,
"learning_rate": 4.495727493049604e-06,
"loss": 0.6361,
"step": 351
},
{
"epoch": 1.594553706505295,
"grad_norm": 1.3316866159439087,
"learning_rate": 4.491843640556033e-06,
"loss": 0.6097,
"step": 352
},
{
"epoch": 1.5990922844175492,
"grad_norm": 1.2030465602874756,
"learning_rate": 4.4879465801224605e-06,
"loss": 0.6302,
"step": 353
},
{
"epoch": 1.6036308623298035,
"grad_norm": 0.9129522442817688,
"learning_rate": 4.484036337590343e-06,
"loss": 0.6398,
"step": 354
},
{
"epoch": 1.6081694402420574,
"grad_norm": 1.6810179948806763,
"learning_rate": 4.4801129388885475e-06,
"loss": 0.6234,
"step": 355
},
{
"epoch": 1.6127080181543116,
"grad_norm": 3.6033570766448975,
"learning_rate": 4.476176410033179e-06,
"loss": 0.6145,
"step": 356
},
{
"epoch": 1.6172465960665658,
"grad_norm": 1.2315465211868286,
"learning_rate": 4.472226777127412e-06,
"loss": 0.6274,
"step": 357
},
{
"epoch": 1.6217851739788198,
"grad_norm": 0.7585744261741638,
"learning_rate": 4.468264066361308e-06,
"loss": 0.5897,
"step": 358
},
{
"epoch": 1.626323751891074,
"grad_norm": 0.945957362651825,
"learning_rate": 4.464288304011652e-06,
"loss": 0.6078,
"step": 359
},
{
"epoch": 1.6308623298033282,
"grad_norm": 1.0154330730438232,
"learning_rate": 4.460299516441777e-06,
"loss": 0.5899,
"step": 360
},
{
"epoch": 1.6354009077155824,
"grad_norm": 0.8923754096031189,
"learning_rate": 4.456297730101379e-06,
"loss": 0.6204,
"step": 361
},
{
"epoch": 1.6399394856278366,
"grad_norm": 0.9550593495368958,
"learning_rate": 4.452282971526355e-06,
"loss": 0.5699,
"step": 362
},
{
"epoch": 1.6444780635400909,
"grad_norm": 0.7987310886383057,
"learning_rate": 4.448255267338619e-06,
"loss": 0.6325,
"step": 363
},
{
"epoch": 1.649016641452345,
"grad_norm": 0.830464780330658,
"learning_rate": 4.444214644245928e-06,
"loss": 0.6367,
"step": 364
},
{
"epoch": 1.653555219364599,
"grad_norm": 1.604446530342102,
"learning_rate": 4.440161129041704e-06,
"loss": 0.5668,
"step": 365
},
{
"epoch": 1.6580937972768532,
"grad_norm": 0.6995673179626465,
"learning_rate": 4.436094748604856e-06,
"loss": 0.596,
"step": 366
},
{
"epoch": 1.6626323751891074,
"grad_norm": 0.6944538950920105,
"learning_rate": 4.432015529899604e-06,
"loss": 0.6082,
"step": 367
},
{
"epoch": 1.6671709531013614,
"grad_norm": 0.9797276854515076,
"learning_rate": 4.427923499975298e-06,
"loss": 0.6104,
"step": 368
},
{
"epoch": 1.6717095310136156,
"grad_norm": 0.7624075412750244,
"learning_rate": 4.423818685966239e-06,
"loss": 0.5721,
"step": 369
},
{
"epoch": 1.6762481089258698,
"grad_norm": 0.8912142515182495,
"learning_rate": 4.4197011150915e-06,
"loss": 0.6162,
"step": 370
},
{
"epoch": 1.680786686838124,
"grad_norm": 1.4948642253875732,
"learning_rate": 4.415570814654746e-06,
"loss": 0.6355,
"step": 371
},
{
"epoch": 1.6853252647503782,
"grad_norm": 3.306320905685425,
"learning_rate": 4.4114278120440494e-06,
"loss": 0.6077,
"step": 372
},
{
"epoch": 1.6898638426626325,
"grad_norm": 0.6849818229675293,
"learning_rate": 4.407272134731711e-06,
"loss": 0.621,
"step": 373
},
{
"epoch": 1.6944024205748867,
"grad_norm": 0.9957187175750732,
"learning_rate": 4.403103810274082e-06,
"loss": 0.6468,
"step": 374
},
{
"epoch": 1.6989409984871406,
"grad_norm": 0.926688551902771,
"learning_rate": 4.398922866311371e-06,
"loss": 0.6021,
"step": 375
},
{
"epoch": 1.7034795763993948,
"grad_norm": 0.8220088481903076,
"learning_rate": 4.394729330567471e-06,
"loss": 0.5753,
"step": 376
},
{
"epoch": 1.708018154311649,
"grad_norm": 0.8064286112785339,
"learning_rate": 4.390523230849769e-06,
"loss": 0.6275,
"step": 377
},
{
"epoch": 1.712556732223903,
"grad_norm": 0.7482770681381226,
"learning_rate": 4.386304595048966e-06,
"loss": 0.6103,
"step": 378
},
{
"epoch": 1.7170953101361572,
"grad_norm": 1.6559797525405884,
"learning_rate": 4.382073451138887e-06,
"loss": 0.6366,
"step": 379
},
{
"epoch": 1.7216338880484114,
"grad_norm": 0.6992952227592468,
"learning_rate": 4.3778298271762995e-06,
"loss": 0.6188,
"step": 380
},
{
"epoch": 1.7261724659606656,
"grad_norm": 0.6812805533409119,
"learning_rate": 4.373573751300729e-06,
"loss": 0.6103,
"step": 381
},
{
"epoch": 1.7307110438729199,
"grad_norm": 0.767241358757019,
"learning_rate": 4.369305251734267e-06,
"loss": 0.6089,
"step": 382
},
{
"epoch": 1.735249621785174,
"grad_norm": 1.5500905513763428,
"learning_rate": 4.365024356781386e-06,
"loss": 0.6087,
"step": 383
},
{
"epoch": 1.7397881996974283,
"grad_norm": 0.8380416631698608,
"learning_rate": 4.360731094828755e-06,
"loss": 0.6074,
"step": 384
},
{
"epoch": 1.7443267776096822,
"grad_norm": 0.915096640586853,
"learning_rate": 4.356425494345047e-06,
"loss": 0.5962,
"step": 385
},
{
"epoch": 1.7488653555219364,
"grad_norm": 0.9544028639793396,
"learning_rate": 4.352107583880753e-06,
"loss": 0.5766,
"step": 386
},
{
"epoch": 1.7534039334341907,
"grad_norm": 0.7770220041275024,
"learning_rate": 4.347777392067991e-06,
"loss": 0.5879,
"step": 387
},
{
"epoch": 1.7579425113464446,
"grad_norm": 3.470493793487549,
"learning_rate": 4.343434947620316e-06,
"loss": 0.6107,
"step": 388
},
{
"epoch": 1.7624810892586988,
"grad_norm": 1.7937536239624023,
"learning_rate": 4.339080279332531e-06,
"loss": 0.5892,
"step": 389
},
{
"epoch": 1.767019667170953,
"grad_norm": 1.262220859527588,
"learning_rate": 4.334713416080498e-06,
"loss": 0.6321,
"step": 390
},
{
"epoch": 1.7715582450832073,
"grad_norm": 0.6722662448883057,
"learning_rate": 4.33033438682094e-06,
"loss": 0.6366,
"step": 391
},
{
"epoch": 1.7760968229954615,
"grad_norm": 1.0483866930007935,
"learning_rate": 4.3259432205912544e-06,
"loss": 0.5867,
"step": 392
},
{
"epoch": 1.7806354009077157,
"grad_norm": 1.2742741107940674,
"learning_rate": 4.32153994650932e-06,
"loss": 0.6045,
"step": 393
},
{
"epoch": 1.7851739788199699,
"grad_norm": 1.2766985893249512,
"learning_rate": 4.317124593773301e-06,
"loss": 0.5952,
"step": 394
},
{
"epoch": 1.789712556732224,
"grad_norm": 0.8167585134506226,
"learning_rate": 4.312697191661457e-06,
"loss": 0.5621,
"step": 395
},
{
"epoch": 1.794251134644478,
"grad_norm": 0.7818560004234314,
"learning_rate": 4.308257769531947e-06,
"loss": 0.5897,
"step": 396
},
{
"epoch": 1.7987897125567323,
"grad_norm": 1.0983150005340576,
"learning_rate": 4.303806356822635e-06,
"loss": 0.6189,
"step": 397
},
{
"epoch": 1.8033282904689862,
"grad_norm": 3.0957119464874268,
"learning_rate": 4.299342983050892e-06,
"loss": 0.5743,
"step": 398
},
{
"epoch": 1.8078668683812404,
"grad_norm": 1.364321231842041,
"learning_rate": 4.294867677813407e-06,
"loss": 0.5722,
"step": 399
},
{
"epoch": 1.8124054462934946,
"grad_norm": 1.0932508707046509,
"learning_rate": 4.290380470785984e-06,
"loss": 0.6074,
"step": 400
},
{
"epoch": 1.8169440242057489,
"grad_norm": 0.8298100829124451,
"learning_rate": 4.285881391723348e-06,
"loss": 0.6143,
"step": 401
},
{
"epoch": 1.821482602118003,
"grad_norm": 0.7821558117866516,
"learning_rate": 4.2813704704589504e-06,
"loss": 0.6148,
"step": 402
},
{
"epoch": 1.8260211800302573,
"grad_norm": 1.2309906482696533,
"learning_rate": 4.276847736904765e-06,
"loss": 0.6039,
"step": 403
},
{
"epoch": 1.8305597579425115,
"grad_norm": 0.7675696015357971,
"learning_rate": 4.272313221051094e-06,
"loss": 0.5869,
"step": 404
},
{
"epoch": 1.8350983358547657,
"grad_norm": 0.6467660069465637,
"learning_rate": 4.267766952966369e-06,
"loss": 0.5999,
"step": 405
},
{
"epoch": 1.8396369137670197,
"grad_norm": 2.9403133392333984,
"learning_rate": 4.263208962796951e-06,
"loss": 0.5859,
"step": 406
},
{
"epoch": 1.8441754916792739,
"grad_norm": 1.017529845237732,
"learning_rate": 4.2586392807669286e-06,
"loss": 0.5771,
"step": 407
},
{
"epoch": 1.8487140695915278,
"grad_norm": 0.7939811944961548,
"learning_rate": 4.25405793717792e-06,
"loss": 0.5968,
"step": 408
},
{
"epoch": 1.853252647503782,
"grad_norm": 0.9015148878097534,
"learning_rate": 4.2494649624088724e-06,
"loss": 0.5791,
"step": 409
},
{
"epoch": 1.8577912254160363,
"grad_norm": 1.0004379749298096,
"learning_rate": 4.2448603869158585e-06,
"loss": 0.5969,
"step": 410
},
{
"epoch": 1.8623298033282905,
"grad_norm": 0.8573418855667114,
"learning_rate": 4.2402442412318765e-06,
"loss": 0.6308,
"step": 411
},
{
"epoch": 1.8668683812405447,
"grad_norm": 0.7616469860076904,
"learning_rate": 4.235616555966646e-06,
"loss": 0.5955,
"step": 412
},
{
"epoch": 1.8714069591527989,
"grad_norm": 0.838377833366394,
"learning_rate": 4.2309773618064035e-06,
"loss": 0.6135,
"step": 413
},
{
"epoch": 1.875945537065053,
"grad_norm": 0.6853629350662231,
"learning_rate": 4.226326689513705e-06,
"loss": 0.5962,
"step": 414
},
{
"epoch": 1.8804841149773073,
"grad_norm": 3.4511594772338867,
"learning_rate": 4.221664569927217e-06,
"loss": 0.632,
"step": 415
},
{
"epoch": 1.8850226928895613,
"grad_norm": 2.316239833831787,
"learning_rate": 4.216991033961511e-06,
"loss": 0.5712,
"step": 416
},
{
"epoch": 1.8895612708018155,
"grad_norm": 0.7679340243339539,
"learning_rate": 4.212306112606863e-06,
"loss": 0.5849,
"step": 417
},
{
"epoch": 1.8940998487140694,
"grad_norm": 0.8144194483757019,
"learning_rate": 4.207609836929045e-06,
"loss": 0.5586,
"step": 418
},
{
"epoch": 1.8986384266263236,
"grad_norm": 0.7225912809371948,
"learning_rate": 4.2029022380691195e-06,
"loss": 0.606,
"step": 419
},
{
"epoch": 1.9031770045385779,
"grad_norm": 0.800234854221344,
"learning_rate": 4.198183347243233e-06,
"loss": 0.6024,
"step": 420
},
{
"epoch": 1.907715582450832,
"grad_norm": 0.7729604840278625,
"learning_rate": 4.1934531957424095e-06,
"loss": 0.598,
"step": 421
},
{
"epoch": 1.9122541603630863,
"grad_norm": 0.6805166602134705,
"learning_rate": 4.188711814932343e-06,
"loss": 0.6148,
"step": 422
},
{
"epoch": 1.9167927382753405,
"grad_norm": 1.0123629570007324,
"learning_rate": 4.1839592362531875e-06,
"loss": 0.6029,
"step": 423
},
{
"epoch": 1.9213313161875947,
"grad_norm": 0.8777531385421753,
"learning_rate": 4.179195491219353e-06,
"loss": 0.5721,
"step": 424
},
{
"epoch": 1.9258698940998489,
"grad_norm": 0.8781999945640564,
"learning_rate": 4.1744206114192895e-06,
"loss": 0.5761,
"step": 425
},
{
"epoch": 1.9304084720121029,
"grad_norm": 0.909726083278656,
"learning_rate": 4.169634628515288e-06,
"loss": 0.6101,
"step": 426
},
{
"epoch": 1.934947049924357,
"grad_norm": 0.8270307779312134,
"learning_rate": 4.164837574243259e-06,
"loss": 0.5635,
"step": 427
},
{
"epoch": 1.939485627836611,
"grad_norm": 0.8078930974006653,
"learning_rate": 4.16002948041253e-06,
"loss": 0.6117,
"step": 428
},
{
"epoch": 1.9440242057488653,
"grad_norm": 1.2975406646728516,
"learning_rate": 4.155210378905629e-06,
"loss": 0.6157,
"step": 429
},
{
"epoch": 1.9485627836611195,
"grad_norm": 1.0585848093032837,
"learning_rate": 4.15038030167808e-06,
"loss": 0.6252,
"step": 430
},
{
"epoch": 1.9531013615733737,
"grad_norm": 1.106614112854004,
"learning_rate": 4.145539280758184e-06,
"loss": 0.5781,
"step": 431
},
{
"epoch": 1.9576399394856279,
"grad_norm": 0.7705745697021484,
"learning_rate": 4.140687348246814e-06,
"loss": 0.6134,
"step": 432
},
{
"epoch": 1.962178517397882,
"grad_norm": 1.1777352094650269,
"learning_rate": 4.1358245363171905e-06,
"loss": 0.6231,
"step": 433
},
{
"epoch": 1.9667170953101363,
"grad_norm": 0.9292231798171997,
"learning_rate": 4.130950877214683e-06,
"loss": 0.6086,
"step": 434
},
{
"epoch": 1.9712556732223905,
"grad_norm": 1.0536510944366455,
"learning_rate": 4.126066403256585e-06,
"loss": 0.6077,
"step": 435
},
{
"epoch": 1.9757942511346445,
"grad_norm": 0.7694706916809082,
"learning_rate": 4.121171146831905e-06,
"loss": 0.6318,
"step": 436
},
{
"epoch": 1.9803328290468987,
"grad_norm": 1.4091219902038574,
"learning_rate": 4.116265140401148e-06,
"loss": 0.5873,
"step": 437
},
{
"epoch": 1.9848714069591527,
"grad_norm": 1.4843878746032715,
"learning_rate": 4.111348416496104e-06,
"loss": 0.5748,
"step": 438
},
{
"epoch": 1.9894099848714069,
"grad_norm": 2.431475877761841,
"learning_rate": 4.106421007719631e-06,
"loss": 0.6155,
"step": 439
},
{
"epoch": 1.993948562783661,
"grad_norm": 0.689834475517273,
"learning_rate": 4.101482946745438e-06,
"loss": 0.5792,
"step": 440
},
{
"epoch": 1.9984871406959153,
"grad_norm": 0.7212426662445068,
"learning_rate": 4.096534266317869e-06,
"loss": 0.6106,
"step": 441
},
{
"epoch": 2.0,
"grad_norm": 0.7212426662445068,
"learning_rate": 4.091574999251685e-06,
"loss": 0.1986,
"step": 442
},
{
"epoch": 2.004538577912254,
"grad_norm": 0.859722912311554,
"learning_rate": 4.086605178431848e-06,
"loss": 0.5424,
"step": 443
},
{
"epoch": 2.0090771558245084,
"grad_norm": 0.782247006893158,
"learning_rate": 4.0816248368133015e-06,
"loss": 0.5169,
"step": 444
},
{
"epoch": 2.0136157337367626,
"grad_norm": 0.7215720415115356,
"learning_rate": 4.076634007420754e-06,
"loss": 0.5712,
"step": 445
},
{
"epoch": 2.018154311649017,
"grad_norm": 0.7926766872406006,
"learning_rate": 4.0716327233484544e-06,
"loss": 0.5433,
"step": 446
},
{
"epoch": 2.022692889561271,
"grad_norm": 0.7854951620101929,
"learning_rate": 4.066621017759984e-06,
"loss": 0.57,
"step": 447
},
{
"epoch": 2.027231467473525,
"grad_norm": 0.7837132215499878,
"learning_rate": 4.0615989238880215e-06,
"loss": 0.5367,
"step": 448
},
{
"epoch": 2.031770045385779,
"grad_norm": 1.0618802309036255,
"learning_rate": 4.056566475034136e-06,
"loss": 0.5247,
"step": 449
},
{
"epoch": 2.036308623298033,
"grad_norm": 0.7492559552192688,
"learning_rate": 4.051523704568557e-06,
"loss": 0.5398,
"step": 450
},
{
"epoch": 2.0408472012102874,
"grad_norm": 1.0183978080749512,
"learning_rate": 4.04647064592996e-06,
"loss": 0.5487,
"step": 451
},
{
"epoch": 2.0453857791225416,
"grad_norm": 0.8974794745445251,
"learning_rate": 4.041407332625238e-06,
"loss": 0.5551,
"step": 452
},
{
"epoch": 2.049924357034796,
"grad_norm": 0.6771326661109924,
"learning_rate": 4.0363337982292865e-06,
"loss": 0.5647,
"step": 453
},
{
"epoch": 2.05446293494705,
"grad_norm": 0.9089450240135193,
"learning_rate": 4.031250076384774e-06,
"loss": 0.5743,
"step": 454
},
{
"epoch": 2.059001512859304,
"grad_norm": 0.641944944858551,
"learning_rate": 4.026156200801924e-06,
"loss": 0.5344,
"step": 455
},
{
"epoch": 2.0635400907715584,
"grad_norm": 0.7394944429397583,
"learning_rate": 4.021052205258288e-06,
"loss": 0.5856,
"step": 456
},
{
"epoch": 2.068078668683812,
"grad_norm": 2.491907835006714,
"learning_rate": 4.015938123598525e-06,
"loss": 0.5597,
"step": 457
},
{
"epoch": 2.0726172465960664,
"grad_norm": 1.3713676929473877,
"learning_rate": 4.010813989734174e-06,
"loss": 0.5324,
"step": 458
},
{
"epoch": 2.0771558245083206,
"grad_norm": 0.7260848879814148,
"learning_rate": 4.00567983764343e-06,
"loss": 0.5529,
"step": 459
},
{
"epoch": 2.081694402420575,
"grad_norm": 1.168824553489685,
"learning_rate": 4.0005357013709215e-06,
"loss": 0.6031,
"step": 460
},
{
"epoch": 2.086232980332829,
"grad_norm": 0.9732924699783325,
"learning_rate": 3.995381615027477e-06,
"loss": 0.5643,
"step": 461
},
{
"epoch": 2.090771558245083,
"grad_norm": 1.2031774520874023,
"learning_rate": 3.990217612789909e-06,
"loss": 0.5651,
"step": 462
},
{
"epoch": 2.0953101361573374,
"grad_norm": 0.7182034850120544,
"learning_rate": 3.985043728900782e-06,
"loss": 0.5507,
"step": 463
},
{
"epoch": 2.0998487140695916,
"grad_norm": 2.1731138229370117,
"learning_rate": 3.979859997668182e-06,
"loss": 0.5581,
"step": 464
},
{
"epoch": 2.104387291981846,
"grad_norm": 1.6525670289993286,
"learning_rate": 3.9746664534654975e-06,
"loss": 0.5827,
"step": 465
},
{
"epoch": 2.1089258698941,
"grad_norm": 0.9597675800323486,
"learning_rate": 3.969463130731183e-06,
"loss": 0.5685,
"step": 466
},
{
"epoch": 2.1134644478063542,
"grad_norm": 0.7419756650924683,
"learning_rate": 3.964250063968537e-06,
"loss": 0.5759,
"step": 467
},
{
"epoch": 2.118003025718608,
"grad_norm": 0.76424241065979,
"learning_rate": 3.959027287745471e-06,
"loss": 0.5648,
"step": 468
},
{
"epoch": 2.122541603630862,
"grad_norm": 0.8645866513252258,
"learning_rate": 3.95379483669428e-06,
"loss": 0.5594,
"step": 469
},
{
"epoch": 2.1270801815431164,
"grad_norm": 1.0087001323699951,
"learning_rate": 3.9485527455114095e-06,
"loss": 0.5359,
"step": 470
},
{
"epoch": 2.1316187594553706,
"grad_norm": 1.055308222770691,
"learning_rate": 3.943301048957233e-06,
"loss": 0.5465,
"step": 471
},
{
"epoch": 2.136157337367625,
"grad_norm": 1.540602445602417,
"learning_rate": 3.9380397818558154e-06,
"loss": 0.5647,
"step": 472
},
{
"epoch": 2.140695915279879,
"grad_norm": 0.6511226892471313,
"learning_rate": 3.932768979094685e-06,
"loss": 0.5102,
"step": 473
},
{
"epoch": 2.145234493192133,
"grad_norm": 0.6932368278503418,
"learning_rate": 3.927488675624599e-06,
"loss": 0.5934,
"step": 474
},
{
"epoch": 2.1497730711043874,
"grad_norm": 0.6622692942619324,
"learning_rate": 3.922198906459318e-06,
"loss": 0.5783,
"step": 475
},
{
"epoch": 2.1543116490166416,
"grad_norm": 0.9521903991699219,
"learning_rate": 3.916899706675366e-06,
"loss": 0.5642,
"step": 476
},
{
"epoch": 2.1588502269288954,
"grad_norm": 1.9041461944580078,
"learning_rate": 3.911591111411802e-06,
"loss": 0.5631,
"step": 477
},
{
"epoch": 2.1633888048411496,
"grad_norm": 0.7302039861679077,
"learning_rate": 3.906273155869988e-06,
"loss": 0.5907,
"step": 478
},
{
"epoch": 2.167927382753404,
"grad_norm": 0.9459813237190247,
"learning_rate": 3.900945875313353e-06,
"loss": 0.5767,
"step": 479
},
{
"epoch": 2.172465960665658,
"grad_norm": 0.9118275046348572,
"learning_rate": 3.895609305067162e-06,
"loss": 0.5699,
"step": 480
},
{
"epoch": 2.177004538577912,
"grad_norm": 0.9915804266929626,
"learning_rate": 3.890263480518278e-06,
"loss": 0.5692,
"step": 481
},
{
"epoch": 2.1815431164901664,
"grad_norm": 0.7956082820892334,
"learning_rate": 3.884908437114931e-06,
"loss": 0.5809,
"step": 482
},
{
"epoch": 2.1860816944024206,
"grad_norm": 0.7874560952186584,
"learning_rate": 3.879544210366479e-06,
"loss": 0.5426,
"step": 483
},
{
"epoch": 2.190620272314675,
"grad_norm": 0.672660231590271,
"learning_rate": 3.8741708358431776e-06,
"loss": 0.5562,
"step": 484
},
{
"epoch": 2.195158850226929,
"grad_norm": 0.9052623510360718,
"learning_rate": 3.868788349175939e-06,
"loss": 0.5946,
"step": 485
},
{
"epoch": 2.1996974281391832,
"grad_norm": 2.023263692855835,
"learning_rate": 3.863396786056102e-06,
"loss": 0.5571,
"step": 486
},
{
"epoch": 2.2042360060514374,
"grad_norm": 0.7047929167747498,
"learning_rate": 3.8579961822351856e-06,
"loss": 0.5659,
"step": 487
},
{
"epoch": 2.208774583963691,
"grad_norm": 0.6957628130912781,
"learning_rate": 3.852586573524663e-06,
"loss": 0.5478,
"step": 488
},
{
"epoch": 2.2133131618759454,
"grad_norm": 1.1220340728759766,
"learning_rate": 3.847167995795716e-06,
"loss": 0.5478,
"step": 489
},
{
"epoch": 2.2178517397881996,
"grad_norm": 1.4658353328704834,
"learning_rate": 3.841740484979002e-06,
"loss": 0.5418,
"step": 490
},
{
"epoch": 2.222390317700454,
"grad_norm": 0.7847384810447693,
"learning_rate": 3.836304077064412e-06,
"loss": 0.5784,
"step": 491
},
{
"epoch": 2.226928895612708,
"grad_norm": 1.5285112857818604,
"learning_rate": 3.830858808100835e-06,
"loss": 0.5449,
"step": 492
},
{
"epoch": 2.231467473524962,
"grad_norm": 0.6902230978012085,
"learning_rate": 3.825404714195917e-06,
"loss": 0.567,
"step": 493
},
{
"epoch": 2.2360060514372164,
"grad_norm": 0.8884925842285156,
"learning_rate": 3.819941831515825e-06,
"loss": 0.5181,
"step": 494
},
{
"epoch": 2.2405446293494706,
"grad_norm": 1.5801842212677002,
"learning_rate": 3.8144701962849973e-06,
"loss": 0.5377,
"step": 495
},
{
"epoch": 2.245083207261725,
"grad_norm": 0.7058039307594299,
"learning_rate": 3.80898984478592e-06,
"loss": 0.5685,
"step": 496
},
{
"epoch": 2.2496217851739786,
"grad_norm": 0.6729607582092285,
"learning_rate": 3.803500813358869e-06,
"loss": 0.563,
"step": 497
},
{
"epoch": 2.254160363086233,
"grad_norm": 0.8975954055786133,
"learning_rate": 3.7980031384016826e-06,
"loss": 0.5865,
"step": 498
},
{
"epoch": 2.258698940998487,
"grad_norm": 1.0153331756591797,
"learning_rate": 3.79249685636951e-06,
"loss": 0.5611,
"step": 499
},
{
"epoch": 2.263237518910741,
"grad_norm": 0.6788516640663147,
"learning_rate": 3.7869820037745773e-06,
"loss": 0.5417,
"step": 500
},
{
"epoch": 2.2677760968229954,
"grad_norm": 0.9200128316879272,
"learning_rate": 3.7814586171859397e-06,
"loss": 0.5621,
"step": 501
},
{
"epoch": 2.2723146747352496,
"grad_norm": 2.8803627490997314,
"learning_rate": 3.775926733229243e-06,
"loss": 0.5593,
"step": 502
},
{
"epoch": 2.276853252647504,
"grad_norm": 2.3636279106140137,
"learning_rate": 3.770386388586479e-06,
"loss": 0.5893,
"step": 503
},
{
"epoch": 2.281391830559758,
"grad_norm": 0.8512002825737,
"learning_rate": 3.7648376199957416e-06,
"loss": 0.5411,
"step": 504
},
{
"epoch": 2.2859304084720122,
"grad_norm": 1.220920205116272,
"learning_rate": 3.7592804642509844e-06,
"loss": 0.5077,
"step": 505
},
{
"epoch": 2.2904689863842664,
"grad_norm": 1.1704233884811401,
"learning_rate": 3.7537149582017764e-06,
"loss": 0.5412,
"step": 506
},
{
"epoch": 2.2950075642965206,
"grad_norm": 1.7733300924301147,
"learning_rate": 3.7481411387530577e-06,
"loss": 0.5297,
"step": 507
},
{
"epoch": 2.2995461422087744,
"grad_norm": 0.6807821393013,
"learning_rate": 3.742559042864895e-06,
"loss": 0.5729,
"step": 508
},
{
"epoch": 2.3040847201210286,
"grad_norm": 1.166459083557129,
"learning_rate": 3.7369687075522355e-06,
"loss": 0.528,
"step": 509
},
{
"epoch": 2.308623298033283,
"grad_norm": 0.7100872993469238,
"learning_rate": 3.7313701698846616e-06,
"loss": 0.5686,
"step": 510
},
{
"epoch": 2.313161875945537,
"grad_norm": 0.9993472099304199,
"learning_rate": 3.725763466986147e-06,
"loss": 0.5676,
"step": 511
},
{
"epoch": 2.317700453857791,
"grad_norm": 0.889721155166626,
"learning_rate": 3.7201486360348075e-06,
"loss": 0.5261,
"step": 512
},
{
"epoch": 2.3222390317700454,
"grad_norm": 0.6611590385437012,
"learning_rate": 3.714525714262659e-06,
"loss": 0.5292,
"step": 513
},
{
"epoch": 2.3267776096822996,
"grad_norm": 0.8129754662513733,
"learning_rate": 3.708894738955364e-06,
"loss": 0.5556,
"step": 514
},
{
"epoch": 2.331316187594554,
"grad_norm": 0.797924816608429,
"learning_rate": 3.703255747451991e-06,
"loss": 0.5294,
"step": 515
},
{
"epoch": 2.335854765506808,
"grad_norm": 0.7335793972015381,
"learning_rate": 3.697608777144762e-06,
"loss": 0.5144,
"step": 516
},
{
"epoch": 2.340393343419062,
"grad_norm": 0.70816570520401,
"learning_rate": 3.691953865478809e-06,
"loss": 0.5649,
"step": 517
},
{
"epoch": 2.344931921331316,
"grad_norm": 0.916517436504364,
"learning_rate": 3.6862910499519204e-06,
"loss": 0.5721,
"step": 518
},
{
"epoch": 2.34947049924357,
"grad_norm": 0.9996108412742615,
"learning_rate": 3.680620368114297e-06,
"loss": 0.5427,
"step": 519
},
{
"epoch": 2.3540090771558244,
"grad_norm": 0.7871035933494568,
"learning_rate": 3.6749418575683005e-06,
"loss": 0.5549,
"step": 520
},
{
"epoch": 2.3585476550680786,
"grad_norm": 0.7144160270690918,
"learning_rate": 3.6692555559682052e-06,
"loss": 0.5779,
"step": 521
},
{
"epoch": 2.363086232980333,
"grad_norm": 0.6424504518508911,
"learning_rate": 3.6635615010199484e-06,
"loss": 0.5392,
"step": 522
},
{
"epoch": 2.367624810892587,
"grad_norm": 1.4431594610214233,
"learning_rate": 3.6578597304808784e-06,
"loss": 0.5398,
"step": 523
},
{
"epoch": 2.3721633888048412,
"grad_norm": 0.834149181842804,
"learning_rate": 3.6521502821595067e-06,
"loss": 0.5421,
"step": 524
},
{
"epoch": 2.3767019667170954,
"grad_norm": 0.802004873752594,
"learning_rate": 3.6464331939152576e-06,
"loss": 0.5086,
"step": 525
},
{
"epoch": 2.3812405446293496,
"grad_norm": 0.73563152551651,
"learning_rate": 3.6407085036582134e-06,
"loss": 0.5406,
"step": 526
},
{
"epoch": 2.385779122541604,
"grad_norm": 1.410982608795166,
"learning_rate": 3.634976249348867e-06,
"loss": 0.568,
"step": 527
},
{
"epoch": 2.3903177004538576,
"grad_norm": 0.8028589487075806,
"learning_rate": 3.629236468997868e-06,
"loss": 0.5085,
"step": 528
},
{
"epoch": 2.394856278366112,
"grad_norm": 0.8530763387680054,
"learning_rate": 3.6234892006657716e-06,
"loss": 0.5324,
"step": 529
},
{
"epoch": 2.399394856278366,
"grad_norm": 0.8291765451431274,
"learning_rate": 3.6177344824627854e-06,
"loss": 0.5593,
"step": 530
},
{
"epoch": 2.40393343419062,
"grad_norm": 0.7214178442955017,
"learning_rate": 3.6119723525485173e-06,
"loss": 0.5302,
"step": 531
},
{
"epoch": 2.4084720121028744,
"grad_norm": 0.9748385548591614,
"learning_rate": 3.606202849131723e-06,
"loss": 0.5362,
"step": 532
},
{
"epoch": 2.4130105900151286,
"grad_norm": 0.8343231678009033,
"learning_rate": 3.600426010470051e-06,
"loss": 0.5762,
"step": 533
},
{
"epoch": 2.417549167927383,
"grad_norm": 1.3052794933319092,
"learning_rate": 3.594641874869792e-06,
"loss": 0.5693,
"step": 534
},
{
"epoch": 2.422087745839637,
"grad_norm": 0.842367947101593,
"learning_rate": 3.5888504806856194e-06,
"loss": 0.5332,
"step": 535
},
{
"epoch": 2.4266263237518912,
"grad_norm": 0.7482140064239502,
"learning_rate": 3.5830518663203412e-06,
"loss": 0.5556,
"step": 536
},
{
"epoch": 2.431164901664145,
"grad_norm": 0.7388250827789307,
"learning_rate": 3.5772460702246415e-06,
"loss": 0.5375,
"step": 537
},
{
"epoch": 2.435703479576399,
"grad_norm": 0.78763747215271,
"learning_rate": 3.5714331308968257e-06,
"loss": 0.5476,
"step": 538
},
{
"epoch": 2.4402420574886534,
"grad_norm": 0.7993074059486389,
"learning_rate": 3.5656130868825677e-06,
"loss": 0.5168,
"step": 539
},
{
"epoch": 2.4447806354009076,
"grad_norm": 0.7094533443450928,
"learning_rate": 3.5597859767746524e-06,
"loss": 0.5483,
"step": 540
},
{
"epoch": 2.449319213313162,
"grad_norm": 0.6727566123008728,
"learning_rate": 3.553951839212718e-06,
"loss": 0.5623,
"step": 541
},
{
"epoch": 2.453857791225416,
"grad_norm": 0.8603148460388184,
"learning_rate": 3.548110712883005e-06,
"loss": 0.5737,
"step": 542
},
{
"epoch": 2.4583963691376702,
"grad_norm": 0.9948770999908447,
"learning_rate": 3.5422626365180936e-06,
"loss": 0.5717,
"step": 543
},
{
"epoch": 2.4629349470499244,
"grad_norm": 0.9731677174568176,
"learning_rate": 3.5364076488966516e-06,
"loss": 0.558,
"step": 544
},
{
"epoch": 2.4674735249621786,
"grad_norm": 1.0563665628433228,
"learning_rate": 3.5305457888431747e-06,
"loss": 0.5143,
"step": 545
},
{
"epoch": 2.472012102874433,
"grad_norm": 0.7944256663322449,
"learning_rate": 3.5246770952277302e-06,
"loss": 0.523,
"step": 546
},
{
"epoch": 2.476550680786687,
"grad_norm": 9.147004127502441,
"learning_rate": 3.5188016069656986e-06,
"loss": 0.5638,
"step": 547
},
{
"epoch": 2.481089258698941,
"grad_norm": 5.153329372406006,
"learning_rate": 3.512919363017516e-06,
"loss": 0.5689,
"step": 548
},
{
"epoch": 2.485627836611195,
"grad_norm": 3.699370861053467,
"learning_rate": 3.5070304023884154e-06,
"loss": 0.5571,
"step": 549
},
{
"epoch": 2.4901664145234492,
"grad_norm": 1.9909933805465698,
"learning_rate": 3.501134764128167e-06,
"loss": 0.5493,
"step": 550
},
{
"epoch": 2.4947049924357034,
"grad_norm": 0.8108904957771301,
"learning_rate": 3.495232487330822e-06,
"loss": 0.5545,
"step": 551
},
{
"epoch": 2.4992435703479576,
"grad_norm": 1.2555493116378784,
"learning_rate": 3.489323611134452e-06,
"loss": 0.5634,
"step": 552
},
{
"epoch": 2.503782148260212,
"grad_norm": 1.1967525482177734,
"learning_rate": 3.4834081747208888e-06,
"loss": 0.5767,
"step": 553
},
{
"epoch": 2.508320726172466,
"grad_norm": 1.1422444581985474,
"learning_rate": 3.477486217315464e-06,
"loss": 0.5774,
"step": 554
},
{
"epoch": 2.5128593040847202,
"grad_norm": 0.7501647472381592,
"learning_rate": 3.4715577781867516e-06,
"loss": 0.5301,
"step": 555
},
{
"epoch": 2.517397881996974,
"grad_norm": 0.6840022802352905,
"learning_rate": 3.465622896646305e-06,
"loss": 0.5548,
"step": 556
},
{
"epoch": 2.521936459909228,
"grad_norm": 0.7143796682357788,
"learning_rate": 3.4596816120483985e-06,
"loss": 0.4968,
"step": 557
},
{
"epoch": 2.5264750378214824,
"grad_norm": 0.7957927584648132,
"learning_rate": 3.453733963789764e-06,
"loss": 0.561,
"step": 558
},
{
"epoch": 2.5310136157337366,
"grad_norm": 2.5668840408325195,
"learning_rate": 3.4477799913093303e-06,
"loss": 0.5766,
"step": 559
},
{
"epoch": 2.535552193645991,
"grad_norm": 0.8086560368537903,
"learning_rate": 3.441819734087963e-06,
"loss": 0.5794,
"step": 560
},
{
"epoch": 2.540090771558245,
"grad_norm": 0.7480419874191284,
"learning_rate": 3.4358532316482037e-06,
"loss": 0.5783,
"step": 561
},
{
"epoch": 2.5446293494704992,
"grad_norm": 0.6926214098930359,
"learning_rate": 3.4298805235540033e-06,
"loss": 0.5488,
"step": 562
},
{
"epoch": 2.5491679273827534,
"grad_norm": 0.6664040088653564,
"learning_rate": 3.4239016494104636e-06,
"loss": 0.5399,
"step": 563
},
{
"epoch": 2.5537065052950076,
"grad_norm": 1.0281089544296265,
"learning_rate": 3.417916648863574e-06,
"loss": 0.5698,
"step": 564
},
{
"epoch": 2.558245083207262,
"grad_norm": 1.0963114500045776,
"learning_rate": 3.411925561599947e-06,
"loss": 0.5392,
"step": 565
},
{
"epoch": 2.562783661119516,
"grad_norm": 0.7198578715324402,
"learning_rate": 3.405928427346557e-06,
"loss": 0.5643,
"step": 566
},
{
"epoch": 2.5673222390317703,
"grad_norm": 0.8547159433364868,
"learning_rate": 3.3999252858704775e-06,
"loss": 0.5077,
"step": 567
},
{
"epoch": 2.5718608169440245,
"grad_norm": 0.7263000011444092,
"learning_rate": 3.3939161769786124e-06,
"loss": 0.496,
"step": 568
},
{
"epoch": 2.5763993948562782,
"grad_norm": 0.6732318997383118,
"learning_rate": 3.387901140517438e-06,
"loss": 0.5575,
"step": 569
},
{
"epoch": 2.5809379727685324,
"grad_norm": 1.9751555919647217,
"learning_rate": 3.3818802163727377e-06,
"loss": 0.5257,
"step": 570
},
{
"epoch": 2.5854765506807866,
"grad_norm": 0.7534604668617249,
"learning_rate": 3.3758534444693323e-06,
"loss": 0.5367,
"step": 571
},
{
"epoch": 2.590015128593041,
"grad_norm": 0.6703462600708008,
"learning_rate": 3.3698208647708226e-06,
"loss": 0.5362,
"step": 572
},
{
"epoch": 2.594553706505295,
"grad_norm": 1.8879098892211914,
"learning_rate": 3.36378251727932e-06,
"loss": 0.5571,
"step": 573
},
{
"epoch": 2.5990922844175492,
"grad_norm": 0.6916099786758423,
"learning_rate": 3.357738442035181e-06,
"loss": 0.5411,
"step": 574
},
{
"epoch": 2.6036308623298035,
"grad_norm": 0.7827025651931763,
"learning_rate": 3.3516886791167446e-06,
"loss": 0.5316,
"step": 575
},
{
"epoch": 2.608169440242057,
"grad_norm": 1.1546365022659302,
"learning_rate": 3.345633268640064e-06,
"loss": 0.5164,
"step": 576
},
{
"epoch": 2.6127080181543114,
"grad_norm": 1.0985122919082642,
"learning_rate": 3.3395722507586413e-06,
"loss": 0.592,
"step": 577
},
{
"epoch": 2.6172465960665656,
"grad_norm": 0.7895015478134155,
"learning_rate": 3.333505665663162e-06,
"loss": 0.5393,
"step": 578
},
{
"epoch": 2.62178517397882,
"grad_norm": 0.8720937967300415,
"learning_rate": 3.327433553581227e-06,
"loss": 0.5474,
"step": 579
},
{
"epoch": 2.626323751891074,
"grad_norm": 0.7074962854385376,
"learning_rate": 3.3213559547770873e-06,
"loss": 0.562,
"step": 580
},
{
"epoch": 2.6308623298033282,
"grad_norm": 1.0457886457443237,
"learning_rate": 3.3152729095513762e-06,
"loss": 0.5188,
"step": 581
},
{
"epoch": 2.6354009077155824,
"grad_norm": 1.1377204656600952,
"learning_rate": 3.309184458240843e-06,
"loss": 0.548,
"step": 582
},
{
"epoch": 2.6399394856278366,
"grad_norm": 1.2311692237854004,
"learning_rate": 3.303090641218083e-06,
"loss": 0.5613,
"step": 583
},
{
"epoch": 2.644478063540091,
"grad_norm": 0.7034773826599121,
"learning_rate": 3.2969914988912746e-06,
"loss": 0.5469,
"step": 584
},
{
"epoch": 2.649016641452345,
"grad_norm": 0.9577970504760742,
"learning_rate": 3.290887071703905e-06,
"loss": 0.5377,
"step": 585
},
{
"epoch": 2.6535552193645993,
"grad_norm": 0.9517496228218079,
"learning_rate": 3.284777400134507e-06,
"loss": 0.5616,
"step": 586
},
{
"epoch": 2.6580937972768535,
"grad_norm": 0.8925998210906982,
"learning_rate": 3.2786625246963903e-06,
"loss": 0.539,
"step": 587
},
{
"epoch": 2.6626323751891077,
"grad_norm": 0.7537391185760498,
"learning_rate": 3.272542485937369e-06,
"loss": 0.5388,
"step": 588
},
{
"epoch": 2.6671709531013614,
"grad_norm": 0.8499334454536438,
"learning_rate": 3.2664173244394965e-06,
"loss": 0.5409,
"step": 589
},
{
"epoch": 2.6717095310136156,
"grad_norm": 1.3763985633850098,
"learning_rate": 3.2602870808187955e-06,
"loss": 0.5425,
"step": 590
},
{
"epoch": 2.67624810892587,
"grad_norm": 0.7515591382980347,
"learning_rate": 3.2541517957249868e-06,
"loss": 0.5623,
"step": 591
},
{
"epoch": 2.680786686838124,
"grad_norm": 0.7507291436195374,
"learning_rate": 3.2480115098412234e-06,
"loss": 0.5335,
"step": 592
},
{
"epoch": 2.6853252647503782,
"grad_norm": 0.7022225856781006,
"learning_rate": 3.2418662638838166e-06,
"loss": 0.5536,
"step": 593
},
{
"epoch": 2.6898638426626325,
"grad_norm": 0.7832311987876892,
"learning_rate": 3.2357160986019697e-06,
"loss": 0.5606,
"step": 594
},
{
"epoch": 2.6944024205748867,
"grad_norm": 1.096700668334961,
"learning_rate": 3.2295610547775054e-06,
"loss": 0.5602,
"step": 595
},
{
"epoch": 2.6989409984871404,
"grad_norm": 0.837489902973175,
"learning_rate": 3.2234011732245953e-06,
"loss": 0.5241,
"step": 596
},
{
"epoch": 2.7034795763993946,
"grad_norm": 0.7516577839851379,
"learning_rate": 3.2172364947894914e-06,
"loss": 0.5557,
"step": 597
},
{
"epoch": 2.708018154311649,
"grad_norm": 1.0745545625686646,
"learning_rate": 3.211067060350253e-06,
"loss": 0.5634,
"step": 598
},
{
"epoch": 2.712556732223903,
"grad_norm": 1.0247219800949097,
"learning_rate": 3.204892910816476e-06,
"loss": 0.5323,
"step": 599
},
{
"epoch": 2.7170953101361572,
"grad_norm": 0.7615066766738892,
"learning_rate": 3.198714087129024e-06,
"loss": 0.5474,
"step": 600
},
{
"epoch": 2.7216338880484114,
"grad_norm": 0.8073952198028564,
"learning_rate": 3.1925306302597535e-06,
"loss": 0.5344,
"step": 601
},
{
"epoch": 2.7261724659606656,
"grad_norm": 1.2883327007293701,
"learning_rate": 3.1863425812112437e-06,
"loss": 0.5876,
"step": 602
},
{
"epoch": 2.73071104387292,
"grad_norm": 2.0846781730651855,
"learning_rate": 3.1801499810165254e-06,
"loss": 0.5452,
"step": 603
},
{
"epoch": 2.735249621785174,
"grad_norm": 0.8865274786949158,
"learning_rate": 3.1739528707388066e-06,
"loss": 0.5386,
"step": 604
},
{
"epoch": 2.7397881996974283,
"grad_norm": 1.5364603996276855,
"learning_rate": 3.1677512914712044e-06,
"loss": 0.5549,
"step": 605
},
{
"epoch": 2.7443267776096825,
"grad_norm": 0.7343050837516785,
"learning_rate": 3.1615452843364674e-06,
"loss": 0.5359,
"step": 606
},
{
"epoch": 2.7488653555219367,
"grad_norm": 0.9987273812294006,
"learning_rate": 3.155334890486707e-06,
"loss": 0.5584,
"step": 607
},
{
"epoch": 2.753403933434191,
"grad_norm": 1.2430927753448486,
"learning_rate": 3.149120151103121e-06,
"loss": 0.5835,
"step": 608
},
{
"epoch": 2.7579425113464446,
"grad_norm": 0.901343047618866,
"learning_rate": 3.142901107395724e-06,
"loss": 0.5629,
"step": 609
},
{
"epoch": 2.762481089258699,
"grad_norm": 1.0777499675750732,
"learning_rate": 3.1366778006030717e-06,
"loss": 0.5638,
"step": 610
},
{
"epoch": 2.767019667170953,
"grad_norm": 0.6995704174041748,
"learning_rate": 3.130450271991991e-06,
"loss": 0.5628,
"step": 611
},
{
"epoch": 2.7715582450832073,
"grad_norm": 0.8021388053894043,
"learning_rate": 3.1242185628573e-06,
"loss": 0.5471,
"step": 612
},
{
"epoch": 2.7760968229954615,
"grad_norm": 1.0073556900024414,
"learning_rate": 3.117982714521541e-06,
"loss": 0.5375,
"step": 613
},
{
"epoch": 2.7806354009077157,
"grad_norm": 1.3955570459365845,
"learning_rate": 3.1117427683347003e-06,
"loss": 0.5481,
"step": 614
},
{
"epoch": 2.78517397881997,
"grad_norm": 0.7922578454017639,
"learning_rate": 3.1054987656739395e-06,
"loss": 0.5338,
"step": 615
},
{
"epoch": 2.789712556732224,
"grad_norm": 0.6501929759979248,
"learning_rate": 3.0992507479433193e-06,
"loss": 0.5076,
"step": 616
},
{
"epoch": 2.794251134644478,
"grad_norm": 1.006299614906311,
"learning_rate": 3.0929987565735214e-06,
"loss": 0.555,
"step": 617
},
{
"epoch": 2.798789712556732,
"grad_norm": 0.7578917145729065,
"learning_rate": 3.0867428330215793e-06,
"loss": 0.5359,
"step": 618
},
{
"epoch": 2.8033282904689862,
"grad_norm": 0.828425943851471,
"learning_rate": 3.0804830187706005e-06,
"loss": 0.5308,
"step": 619
},
{
"epoch": 2.8078668683812404,
"grad_norm": 0.7022498846054077,
"learning_rate": 3.0742193553294896e-06,
"loss": 0.5861,
"step": 620
},
{
"epoch": 2.8124054462934946,
"grad_norm": 0.6671915054321289,
"learning_rate": 3.067951884232678e-06,
"loss": 0.5764,
"step": 621
},
{
"epoch": 2.816944024205749,
"grad_norm": 0.9621557593345642,
"learning_rate": 3.0616806470398453e-06,
"loss": 0.5567,
"step": 622
},
{
"epoch": 2.821482602118003,
"grad_norm": 0.7543350458145142,
"learning_rate": 3.055405685335643e-06,
"loss": 0.5249,
"step": 623
},
{
"epoch": 2.8260211800302573,
"grad_norm": 0.8421291708946228,
"learning_rate": 3.0491270407294195e-06,
"loss": 0.5657,
"step": 624
},
{
"epoch": 2.8305597579425115,
"grad_norm": 0.7241919636726379,
"learning_rate": 3.0428447548549466e-06,
"loss": 0.5679,
"step": 625
},
{
"epoch": 2.8350983358547657,
"grad_norm": 0.7716313004493713,
"learning_rate": 3.03655886937014e-06,
"loss": 0.562,
"step": 626
},
{
"epoch": 2.83963691376702,
"grad_norm": 2.4703383445739746,
"learning_rate": 3.030269425956784e-06,
"loss": 0.5356,
"step": 627
},
{
"epoch": 2.844175491679274,
"grad_norm": 1.180924415588379,
"learning_rate": 3.0239764663202565e-06,
"loss": 0.5619,
"step": 628
},
{
"epoch": 2.848714069591528,
"grad_norm": 0.7013900876045227,
"learning_rate": 3.017680032189252e-06,
"loss": 0.5415,
"step": 629
},
{
"epoch": 2.853252647503782,
"grad_norm": 1.6009024381637573,
"learning_rate": 3.011380165315503e-06,
"loss": 0.5618,
"step": 630
},
{
"epoch": 2.8577912254160363,
"grad_norm": 1.0912448167800903,
"learning_rate": 3.005076907473505e-06,
"loss": 0.5821,
"step": 631
},
{
"epoch": 2.8623298033282905,
"grad_norm": 0.8813812732696533,
"learning_rate": 2.9987703004602394e-06,
"loss": 0.5405,
"step": 632
},
{
"epoch": 2.8668683812405447,
"grad_norm": 0.8485713601112366,
"learning_rate": 2.9924603860948963e-06,
"loss": 0.5189,
"step": 633
},
{
"epoch": 2.871406959152799,
"grad_norm": 0.6998408436775208,
"learning_rate": 2.986147206218597e-06,
"loss": 0.5483,
"step": 634
},
{
"epoch": 2.875945537065053,
"grad_norm": 0.7725822925567627,
"learning_rate": 2.9798308026941147e-06,
"loss": 0.5553,
"step": 635
},
{
"epoch": 2.8804841149773073,
"grad_norm": 1.0571022033691406,
"learning_rate": 2.973511217405601e-06,
"loss": 0.5628,
"step": 636
},
{
"epoch": 2.885022692889561,
"grad_norm": 2.2442429065704346,
"learning_rate": 2.967188492258304e-06,
"loss": 0.567,
"step": 637
},
{
"epoch": 2.8895612708018152,
"grad_norm": 1.3726474046707153,
"learning_rate": 2.9608626691782927e-06,
"loss": 0.546,
"step": 638
},
{
"epoch": 2.8940998487140694,
"grad_norm": 0.6917457580566406,
"learning_rate": 2.9545337901121796e-06,
"loss": 0.5322,
"step": 639
},
{
"epoch": 2.8986384266263236,
"grad_norm": 0.8719446659088135,
"learning_rate": 2.9482018970268395e-06,
"loss": 0.4992,
"step": 640
},
{
"epoch": 2.903177004538578,
"grad_norm": 0.8265273571014404,
"learning_rate": 2.941867031909136e-06,
"loss": 0.5641,
"step": 641
},
{
"epoch": 2.907715582450832,
"grad_norm": 0.9295107126235962,
"learning_rate": 2.9355292367656363e-06,
"loss": 0.5269,
"step": 642
},
{
"epoch": 2.9122541603630863,
"grad_norm": 1.0216796398162842,
"learning_rate": 2.9291885536223415e-06,
"loss": 0.543,
"step": 643
},
{
"epoch": 2.9167927382753405,
"grad_norm": 0.7285480499267578,
"learning_rate": 2.9228450245243994e-06,
"loss": 0.5021,
"step": 644
},
{
"epoch": 2.9213313161875947,
"grad_norm": 1.5654610395431519,
"learning_rate": 2.91649869153583e-06,
"loss": 0.5414,
"step": 645
},
{
"epoch": 2.925869894099849,
"grad_norm": 0.9667910933494568,
"learning_rate": 2.910149596739248e-06,
"loss": 0.5309,
"step": 646
},
{
"epoch": 2.930408472012103,
"grad_norm": 0.7517403364181519,
"learning_rate": 2.9037977822355783e-06,
"loss": 0.5385,
"step": 647
},
{
"epoch": 2.9349470499243573,
"grad_norm": 0.9054082036018372,
"learning_rate": 2.8974432901437827e-06,
"loss": 0.5404,
"step": 648
},
{
"epoch": 2.939485627836611,
"grad_norm": 0.8233144879341125,
"learning_rate": 2.8910861626005774e-06,
"loss": 0.5533,
"step": 649
},
{
"epoch": 2.9440242057488653,
"grad_norm": 0.7188256978988647,
"learning_rate": 2.884726441760155e-06,
"loss": 0.5448,
"step": 650
},
{
"epoch": 2.9485627836611195,
"grad_norm": 0.8374635577201843,
"learning_rate": 2.878364169793903e-06,
"loss": 0.5596,
"step": 651
},
{
"epoch": 2.9531013615733737,
"grad_norm": 1.4161272048950195,
"learning_rate": 2.871999388890126e-06,
"loss": 0.5863,
"step": 652
},
{
"epoch": 2.957639939485628,
"grad_norm": 1.0405840873718262,
"learning_rate": 2.8656321412537653e-06,
"loss": 0.5196,
"step": 653
},
{
"epoch": 2.962178517397882,
"grad_norm": 0.9523102641105652,
"learning_rate": 2.85926246910612e-06,
"loss": 0.5287,
"step": 654
},
{
"epoch": 2.9667170953101363,
"grad_norm": 1.1815980672836304,
"learning_rate": 2.8528904146845652e-06,
"loss": 0.5453,
"step": 655
},
{
"epoch": 2.9712556732223905,
"grad_norm": 2.321892499923706,
"learning_rate": 2.8465160202422737e-06,
"loss": 0.5703,
"step": 656
},
{
"epoch": 2.9757942511346442,
"grad_norm": 0.9072638750076294,
"learning_rate": 2.840139328047934e-06,
"loss": 0.5634,
"step": 657
},
{
"epoch": 2.9803328290468984,
"grad_norm": 0.8406242728233337,
"learning_rate": 2.8337603803854713e-06,
"loss": 0.5464,
"step": 658
},
{
"epoch": 2.9848714069591527,
"grad_norm": 0.9542201161384583,
"learning_rate": 2.8273792195537663e-06,
"loss": 0.5129,
"step": 659
},
{
"epoch": 2.989409984871407,
"grad_norm": 0.9052215814590454,
"learning_rate": 2.820995887866378e-06,
"loss": 0.5462,
"step": 660
},
{
"epoch": 2.993948562783661,
"grad_norm": 1.0007253885269165,
"learning_rate": 2.8146104276512565e-06,
"loss": 0.555,
"step": 661
},
{
"epoch": 2.9984871406959153,
"grad_norm": 0.9045431613922119,
"learning_rate": 2.8082228812504693e-06,
"loss": 0.5542,
"step": 662
},
{
"epoch": 3.0,
"grad_norm": 0.7561860084533691,
"learning_rate": 2.801833291019915e-06,
"loss": 0.1716,
"step": 663
},
{
"epoch": 3.004538577912254,
"grad_norm": 1.0536010265350342,
"learning_rate": 2.7954416993290474e-06,
"loss": 0.5426,
"step": 664
},
{
"epoch": 3.0090771558245084,
"grad_norm": 1.148224949836731,
"learning_rate": 2.7890481485605898e-06,
"loss": 0.531,
"step": 665
},
{
"epoch": 3.0136157337367626,
"grad_norm": 1.2489687204360962,
"learning_rate": 2.7826526811102577e-06,
"loss": 0.5283,
"step": 666
},
{
"epoch": 3.018154311649017,
"grad_norm": 0.7956582903862,
"learning_rate": 2.7762553393864743e-06,
"loss": 0.5213,
"step": 667
},
{
"epoch": 3.022692889561271,
"grad_norm": 1.8829700946807861,
"learning_rate": 2.769856165810093e-06,
"loss": 0.5051,
"step": 668
},
{
"epoch": 3.027231467473525,
"grad_norm": 0.6515493392944336,
"learning_rate": 2.7634552028141137e-06,
"loss": 0.5136,
"step": 669
},
{
"epoch": 3.031770045385779,
"grad_norm": 0.8646936416625977,
"learning_rate": 2.757052492843401e-06,
"loss": 0.5044,
"step": 670
},
{
"epoch": 3.036308623298033,
"grad_norm": 0.7098562717437744,
"learning_rate": 2.750648078354406e-06,
"loss": 0.5072,
"step": 671
},
{
"epoch": 3.0408472012102874,
"grad_norm": 0.7068854570388794,
"learning_rate": 2.7442420018148797e-06,
"loss": 0.4857,
"step": 672
},
{
"epoch": 3.0453857791225416,
"grad_norm": 0.8640517592430115,
"learning_rate": 2.7378343057035956e-06,
"loss": 0.5595,
"step": 673
},
{
"epoch": 3.049924357034796,
"grad_norm": 0.6892681121826172,
"learning_rate": 2.7314250325100667e-06,
"loss": 0.5237,
"step": 674
},
{
"epoch": 3.05446293494705,
"grad_norm": 0.6955535411834717,
"learning_rate": 2.7250142247342637e-06,
"loss": 0.5219,
"step": 675
},
{
"epoch": 3.059001512859304,
"grad_norm": 0.8527218103408813,
"learning_rate": 2.718601924886332e-06,
"loss": 0.5392,
"step": 676
},
{
"epoch": 3.0635400907715584,
"grad_norm": 0.7058966159820557,
"learning_rate": 2.7121881754863126e-06,
"loss": 0.5035,
"step": 677
},
{
"epoch": 3.068078668683812,
"grad_norm": 1.1376885175704956,
"learning_rate": 2.7057730190638575e-06,
"loss": 0.4946,
"step": 678
},
{
"epoch": 3.0726172465960664,
"grad_norm": 0.8086661100387573,
"learning_rate": 2.699356498157949e-06,
"loss": 0.4892,
"step": 679
},
{
"epoch": 3.0771558245083206,
"grad_norm": 0.6907851696014404,
"learning_rate": 2.6929386553166165e-06,
"loss": 0.5366,
"step": 680
},
{
"epoch": 3.081694402420575,
"grad_norm": 0.6702793836593628,
"learning_rate": 2.686519533096656e-06,
"loss": 0.497,
"step": 681
},
{
"epoch": 3.086232980332829,
"grad_norm": 1.3654205799102783,
"learning_rate": 2.680099174063348e-06,
"loss": 0.5224,
"step": 682
},
{
"epoch": 3.090771558245083,
"grad_norm": 0.7771999835968018,
"learning_rate": 2.673677620790172e-06,
"loss": 0.5089,
"step": 683
},
{
"epoch": 3.0953101361573374,
"grad_norm": 0.8415461182594299,
"learning_rate": 2.667254915858529e-06,
"loss": 0.5286,
"step": 684
},
{
"epoch": 3.0998487140695916,
"grad_norm": 0.8293631076812744,
"learning_rate": 2.6608311018574545e-06,
"loss": 0.4751,
"step": 685
},
{
"epoch": 3.104387291981846,
"grad_norm": 1.1603025197982788,
"learning_rate": 2.6544062213833395e-06,
"loss": 0.486,
"step": 686
},
{
"epoch": 3.1089258698941,
"grad_norm": 0.7829399704933167,
"learning_rate": 2.647980317039646e-06,
"loss": 0.5243,
"step": 687
},
{
"epoch": 3.1134644478063542,
"grad_norm": 0.6867077946662903,
"learning_rate": 2.6415534314366264e-06,
"loss": 0.519,
"step": 688
},
{
"epoch": 3.118003025718608,
"grad_norm": 0.8667089939117432,
"learning_rate": 2.635125607191039e-06,
"loss": 0.4992,
"step": 689
},
{
"epoch": 3.122541603630862,
"grad_norm": 0.8637856841087341,
"learning_rate": 2.6286968869258666e-06,
"loss": 0.5129,
"step": 690
},
{
"epoch": 3.1270801815431164,
"grad_norm": 0.6830906867980957,
"learning_rate": 2.6222673132700335e-06,
"loss": 0.524,
"step": 691
},
{
"epoch": 3.1316187594553706,
"grad_norm": 0.8948183059692383,
"learning_rate": 2.615836928858122e-06,
"loss": 0.5153,
"step": 692
},
{
"epoch": 3.136157337367625,
"grad_norm": 0.7270144820213318,
"learning_rate": 2.609405776330092e-06,
"loss": 0.5076,
"step": 693
},
{
"epoch": 3.140695915279879,
"grad_norm": 0.6997873783111572,
"learning_rate": 2.6029738983309954e-06,
"loss": 0.5008,
"step": 694
},
{
"epoch": 3.145234493192133,
"grad_norm": 0.7153282165527344,
"learning_rate": 2.5965413375106965e-06,
"loss": 0.5356,
"step": 695
},
{
"epoch": 3.1497730711043874,
"grad_norm": 0.8726050853729248,
"learning_rate": 2.5901081365235852e-06,
"loss": 0.5031,
"step": 696
},
{
"epoch": 3.1543116490166416,
"grad_norm": 0.9993102550506592,
"learning_rate": 2.583674338028298e-06,
"loss": 0.5487,
"step": 697
},
{
"epoch": 3.1588502269288954,
"grad_norm": 1.1097780466079712,
"learning_rate": 2.5772399846874323e-06,
"loss": 0.5031,
"step": 698
},
{
"epoch": 3.1633888048411496,
"grad_norm": 0.876848578453064,
"learning_rate": 2.5708051191672658e-06,
"loss": 0.5504,
"step": 699
},
{
"epoch": 3.167927382753404,
"grad_norm": 0.7725114226341248,
"learning_rate": 2.5643697841374722e-06,
"loss": 0.48,
"step": 700
},
{
"epoch": 3.172465960665658,
"grad_norm": 0.9931405186653137,
"learning_rate": 2.557934022270837e-06,
"loss": 0.5078,
"step": 701
},
{
"epoch": 3.177004538577912,
"grad_norm": 0.8958150744438171,
"learning_rate": 2.551497876242978e-06,
"loss": 0.5278,
"step": 702
},
{
"epoch": 3.1815431164901664,
"grad_norm": 1.0840779542922974,
"learning_rate": 2.5450613887320606e-06,
"loss": 0.5126,
"step": 703
},
{
"epoch": 3.1860816944024206,
"grad_norm": 0.6236334443092346,
"learning_rate": 2.538624602418513e-06,
"loss": 0.4677,
"step": 704
},
{
"epoch": 3.190620272314675,
"grad_norm": 0.731157660484314,
"learning_rate": 2.5321875599847456e-06,
"loss": 0.5218,
"step": 705
},
{
"epoch": 3.195158850226929,
"grad_norm": 0.8157845139503479,
"learning_rate": 2.525750304114867e-06,
"loss": 0.5423,
"step": 706
},
{
"epoch": 3.1996974281391832,
"grad_norm": 1.5604270696640015,
"learning_rate": 2.519312877494401e-06,
"loss": 0.4742,
"step": 707
},
{
"epoch": 3.2042360060514374,
"grad_norm": 0.7842020988464355,
"learning_rate": 2.512875322810002e-06,
"loss": 0.5113,
"step": 708
},
{
"epoch": 3.208774583963691,
"grad_norm": 0.8386942744255066,
"learning_rate": 2.5064376827491786e-06,
"loss": 0.4777,
"step": 709
},
{
"epoch": 3.2133131618759454,
"grad_norm": 2.208853244781494,
"learning_rate": 2.5e-06,
"loss": 0.5465,
"step": 710
},
{
"epoch": 3.2178517397881996,
"grad_norm": 0.8791889548301697,
"learning_rate": 2.4935623172508223e-06,
"loss": 0.5285,
"step": 711
},
{
"epoch": 3.222390317700454,
"grad_norm": 0.7482249140739441,
"learning_rate": 2.4871246771899983e-06,
"loss": 0.5327,
"step": 712
},
{
"epoch": 3.226928895612708,
"grad_norm": 0.6846868395805359,
"learning_rate": 2.4806871225056006e-06,
"loss": 0.5253,
"step": 713
},
{
"epoch": 3.231467473524962,
"grad_norm": 0.7808830738067627,
"learning_rate": 2.474249695885134e-06,
"loss": 0.5438,
"step": 714
},
{
"epoch": 3.2360060514372164,
"grad_norm": 0.7752459645271301,
"learning_rate": 2.467812440015255e-06,
"loss": 0.5232,
"step": 715
},
{
"epoch": 3.2405446293494706,
"grad_norm": 0.7170999646186829,
"learning_rate": 2.461375397581487e-06,
"loss": 0.4919,
"step": 716
},
{
"epoch": 3.245083207261725,
"grad_norm": 0.7586008906364441,
"learning_rate": 2.4549386112679394e-06,
"loss": 0.5218,
"step": 717
},
{
"epoch": 3.2496217851739786,
"grad_norm": 0.7302015423774719,
"learning_rate": 2.448502123757022e-06,
"loss": 0.5271,
"step": 718
},
{
"epoch": 3.254160363086233,
"grad_norm": 0.7475630044937134,
"learning_rate": 2.4420659777291637e-06,
"loss": 0.5189,
"step": 719
},
{
"epoch": 3.258698940998487,
"grad_norm": 0.6879778504371643,
"learning_rate": 2.435630215862529e-06,
"loss": 0.5322,
"step": 720
},
{
"epoch": 3.263237518910741,
"grad_norm": 0.7178695797920227,
"learning_rate": 2.4291948808327346e-06,
"loss": 0.511,
"step": 721
},
{
"epoch": 3.2677760968229954,
"grad_norm": 0.9109921455383301,
"learning_rate": 2.422760015312568e-06,
"loss": 0.4752,
"step": 722
},
{
"epoch": 3.2723146747352496,
"grad_norm": 0.6712206602096558,
"learning_rate": 2.416325661971703e-06,
"loss": 0.5074,
"step": 723
},
{
"epoch": 3.276853252647504,
"grad_norm": 1.5459402799606323,
"learning_rate": 2.4098918634764156e-06,
"loss": 0.502,
"step": 724
},
{
"epoch": 3.281391830559758,
"grad_norm": 0.7668531537055969,
"learning_rate": 2.403458662489304e-06,
"loss": 0.5012,
"step": 725
},
{
"epoch": 3.2859304084720122,
"grad_norm": 0.7114011645317078,
"learning_rate": 2.397026101669005e-06,
"loss": 0.5506,
"step": 726
},
{
"epoch": 3.2904689863842664,
"grad_norm": 0.749311625957489,
"learning_rate": 2.3905942236699086e-06,
"loss": 0.5321,
"step": 727
},
{
"epoch": 3.2950075642965206,
"grad_norm": 0.7871769070625305,
"learning_rate": 2.3841630711418784e-06,
"loss": 0.4805,
"step": 728
},
{
"epoch": 3.2995461422087744,
"grad_norm": 0.7565982937812805,
"learning_rate": 2.377732686729967e-06,
"loss": 0.5431,
"step": 729
},
{
"epoch": 3.3040847201210286,
"grad_norm": 0.8726269006729126,
"learning_rate": 2.371303113074134e-06,
"loss": 0.5153,
"step": 730
},
{
"epoch": 3.308623298033283,
"grad_norm": 0.7536921501159668,
"learning_rate": 2.3648743928089612e-06,
"loss": 0.4863,
"step": 731
},
{
"epoch": 3.313161875945537,
"grad_norm": 0.796808660030365,
"learning_rate": 2.358446568563374e-06,
"loss": 0.4908,
"step": 732
},
{
"epoch": 3.317700453857791,
"grad_norm": 0.7959126830101013,
"learning_rate": 2.3520196829603547e-06,
"loss": 0.5232,
"step": 733
},
{
"epoch": 3.3222390317700454,
"grad_norm": 1.1234790086746216,
"learning_rate": 2.3455937786166613e-06,
"loss": 0.5304,
"step": 734
},
{
"epoch": 3.3267776096822996,
"grad_norm": 0.687144935131073,
"learning_rate": 2.3391688981425464e-06,
"loss": 0.5027,
"step": 735
},
{
"epoch": 3.331316187594554,
"grad_norm": 0.7692601680755615,
"learning_rate": 2.3327450841414716e-06,
"loss": 0.5021,
"step": 736
},
{
"epoch": 3.335854765506808,
"grad_norm": 0.7073154449462891,
"learning_rate": 2.3263223792098287e-06,
"loss": 0.536,
"step": 737
},
{
"epoch": 3.340393343419062,
"grad_norm": 0.9052091240882874,
"learning_rate": 2.3199008259366524e-06,
"loss": 0.5473,
"step": 738
},
{
"epoch": 3.344931921331316,
"grad_norm": 0.8449652791023254,
"learning_rate": 2.3134804669033437e-06,
"loss": 0.5151,
"step": 739
},
{
"epoch": 3.34947049924357,
"grad_norm": 0.7386903166770935,
"learning_rate": 2.3070613446833843e-06,
"loss": 0.5218,
"step": 740
},
{
"epoch": 3.3540090771558244,
"grad_norm": 1.0674140453338623,
"learning_rate": 2.300643501842052e-06,
"loss": 0.4955,
"step": 741
},
{
"epoch": 3.3585476550680786,
"grad_norm": 0.6917946338653564,
"learning_rate": 2.294226980936143e-06,
"loss": 0.5357,
"step": 742
},
{
"epoch": 3.363086232980333,
"grad_norm": 0.735014021396637,
"learning_rate": 2.287811824513688e-06,
"loss": 0.5045,
"step": 743
},
{
"epoch": 3.367624810892587,
"grad_norm": 0.9542885422706604,
"learning_rate": 2.2813980751136686e-06,
"loss": 0.52,
"step": 744
},
{
"epoch": 3.3721633888048412,
"grad_norm": 0.8511250019073486,
"learning_rate": 2.274985775265737e-06,
"loss": 0.5107,
"step": 745
},
{
"epoch": 3.3767019667170954,
"grad_norm": 0.7510440945625305,
"learning_rate": 2.2685749674899346e-06,
"loss": 0.4872,
"step": 746
},
{
"epoch": 3.3812405446293496,
"grad_norm": 0.7399076819419861,
"learning_rate": 2.262165694296406e-06,
"loss": 0.5187,
"step": 747
},
{
"epoch": 3.385779122541604,
"grad_norm": 0.6950759887695312,
"learning_rate": 2.255757998185122e-06,
"loss": 0.5205,
"step": 748
},
{
"epoch": 3.3903177004538576,
"grad_norm": 0.736414909362793,
"learning_rate": 2.2493519216455945e-06,
"loss": 0.5092,
"step": 749
},
{
"epoch": 3.394856278366112,
"grad_norm": 1.004551649093628,
"learning_rate": 2.242947507156599e-06,
"loss": 0.5326,
"step": 750
},
{
"epoch": 3.399394856278366,
"grad_norm": 3.4429965019226074,
"learning_rate": 2.2365447971858868e-06,
"loss": 0.547,
"step": 751
},
{
"epoch": 3.40393343419062,
"grad_norm": 0.672024130821228,
"learning_rate": 2.2301438341899073e-06,
"loss": 0.4979,
"step": 752
},
{
"epoch": 3.4084720121028744,
"grad_norm": 0.8082073926925659,
"learning_rate": 2.223744660613526e-06,
"loss": 0.5099,
"step": 753
},
{
"epoch": 3.4130105900151286,
"grad_norm": 0.7335776090621948,
"learning_rate": 2.217347318889743e-06,
"loss": 0.5196,
"step": 754
},
{
"epoch": 3.417549167927383,
"grad_norm": 0.791305422782898,
"learning_rate": 2.210951851439411e-06,
"loss": 0.5072,
"step": 755
},
{
"epoch": 3.422087745839637,
"grad_norm": 0.8634244203567505,
"learning_rate": 2.204558300670954e-06,
"loss": 0.5095,
"step": 756
},
{
"epoch": 3.4266263237518912,
"grad_norm": 0.910862147808075,
"learning_rate": 2.198166708980086e-06,
"loss": 0.5334,
"step": 757
},
{
"epoch": 3.431164901664145,
"grad_norm": 0.685192346572876,
"learning_rate": 2.191777118749532e-06,
"loss": 0.5061,
"step": 758
},
{
"epoch": 3.435703479576399,
"grad_norm": 0.7736103534698486,
"learning_rate": 2.185389572348745e-06,
"loss": 0.5209,
"step": 759
},
{
"epoch": 3.4402420574886534,
"grad_norm": 2.14925217628479,
"learning_rate": 2.1790041121336223e-06,
"loss": 0.5325,
"step": 760
},
{
"epoch": 3.4447806354009076,
"grad_norm": 0.7848958969116211,
"learning_rate": 2.1726207804462336e-06,
"loss": 0.5217,
"step": 761
},
{
"epoch": 3.449319213313162,
"grad_norm": 0.8946971893310547,
"learning_rate": 2.1662396196145295e-06,
"loss": 0.5028,
"step": 762
},
{
"epoch": 3.453857791225416,
"grad_norm": 0.7946920394897461,
"learning_rate": 2.1598606719520663e-06,
"loss": 0.4991,
"step": 763
},
{
"epoch": 3.4583963691376702,
"grad_norm": 0.707175076007843,
"learning_rate": 2.153483979757727e-06,
"loss": 0.5172,
"step": 764
},
{
"epoch": 3.4629349470499244,
"grad_norm": 0.8119880557060242,
"learning_rate": 2.147109585315435e-06,
"loss": 0.4985,
"step": 765
},
{
"epoch": 3.4674735249621786,
"grad_norm": 0.8942325711250305,
"learning_rate": 2.1407375308938807e-06,
"loss": 0.4909,
"step": 766
},
{
"epoch": 3.472012102874433,
"grad_norm": 0.9520535469055176,
"learning_rate": 2.134367858746236e-06,
"loss": 0.5228,
"step": 767
},
{
"epoch": 3.476550680786687,
"grad_norm": 0.9031659364700317,
"learning_rate": 2.1280006111098754e-06,
"loss": 0.5083,
"step": 768
},
{
"epoch": 3.481089258698941,
"grad_norm": 0.9307310581207275,
"learning_rate": 2.1216358302060987e-06,
"loss": 0.5067,
"step": 769
},
{
"epoch": 3.485627836611195,
"grad_norm": 0.7787923812866211,
"learning_rate": 2.1152735582398453e-06,
"loss": 0.4929,
"step": 770
},
{
"epoch": 3.4901664145234492,
"grad_norm": 0.7476038932800293,
"learning_rate": 2.1089138373994226e-06,
"loss": 0.5082,
"step": 771
},
{
"epoch": 3.4947049924357034,
"grad_norm": 1.7294312715530396,
"learning_rate": 2.1025567098562177e-06,
"loss": 0.511,
"step": 772
},
{
"epoch": 3.4992435703479576,
"grad_norm": 1.9948171377182007,
"learning_rate": 2.096202217764422e-06,
"loss": 0.5264,
"step": 773
},
{
"epoch": 3.503782148260212,
"grad_norm": 0.7554813623428345,
"learning_rate": 2.089850403260753e-06,
"loss": 0.4951,
"step": 774
},
{
"epoch": 3.508320726172466,
"grad_norm": 0.8657441139221191,
"learning_rate": 2.0835013084641704e-06,
"loss": 0.4973,
"step": 775
},
{
"epoch": 3.5128593040847202,
"grad_norm": 1.0366365909576416,
"learning_rate": 2.0771549754756014e-06,
"loss": 0.5198,
"step": 776
},
{
"epoch": 3.517397881996974,
"grad_norm": 0.9612494707107544,
"learning_rate": 2.070811446377659e-06,
"loss": 0.4916,
"step": 777
},
{
"epoch": 3.521936459909228,
"grad_norm": 0.7001491785049438,
"learning_rate": 2.064470763234364e-06,
"loss": 0.5143,
"step": 778
},
{
"epoch": 3.5264750378214824,
"grad_norm": 0.7265450954437256,
"learning_rate": 2.0581329680908654e-06,
"loss": 0.4852,
"step": 779
},
{
"epoch": 3.5310136157337366,
"grad_norm": 0.7637912631034851,
"learning_rate": 2.0517981029731613e-06,
"loss": 0.5101,
"step": 780
},
{
"epoch": 3.535552193645991,
"grad_norm": 0.7350311279296875,
"learning_rate": 2.045466209887821e-06,
"loss": 0.4884,
"step": 781
},
{
"epoch": 3.540090771558245,
"grad_norm": 2.7524001598358154,
"learning_rate": 2.0391373308217077e-06,
"loss": 0.5007,
"step": 782
},
{
"epoch": 3.5446293494704992,
"grad_norm": 0.773992657661438,
"learning_rate": 2.032811507741697e-06,
"loss": 0.5144,
"step": 783
},
{
"epoch": 3.5491679273827534,
"grad_norm": 0.7631218433380127,
"learning_rate": 2.0264887825944e-06,
"loss": 0.5249,
"step": 784
},
{
"epoch": 3.5537065052950076,
"grad_norm": 0.8911067843437195,
"learning_rate": 2.020169197305886e-06,
"loss": 0.5218,
"step": 785
},
{
"epoch": 3.558245083207262,
"grad_norm": 0.6976955533027649,
"learning_rate": 2.013852793781404e-06,
"loss": 0.4844,
"step": 786
},
{
"epoch": 3.562783661119516,
"grad_norm": 0.768477201461792,
"learning_rate": 2.007539613905104e-06,
"loss": 0.5026,
"step": 787
},
{
"epoch": 3.5673222390317703,
"grad_norm": 0.6874233484268188,
"learning_rate": 2.0012296995397614e-06,
"loss": 0.5416,
"step": 788
},
{
"epoch": 3.5718608169440245,
"grad_norm": 0.9146207571029663,
"learning_rate": 1.9949230925264963e-06,
"loss": 0.5285,
"step": 789
},
{
"epoch": 3.5763993948562782,
"grad_norm": 1.090366244316101,
"learning_rate": 1.988619834684499e-06,
"loss": 0.5133,
"step": 790
},
{
"epoch": 3.5809379727685324,
"grad_norm": 0.9968334436416626,
"learning_rate": 1.982319967810749e-06,
"loss": 0.4654,
"step": 791
},
{
"epoch": 3.5854765506807866,
"grad_norm": 0.7116790413856506,
"learning_rate": 1.976023533679744e-06,
"loss": 0.5077,
"step": 792
},
{
"epoch": 3.590015128593041,
"grad_norm": 1.0453625917434692,
"learning_rate": 1.969730574043217e-06,
"loss": 0.5124,
"step": 793
},
{
"epoch": 3.594553706505295,
"grad_norm": 0.7500141859054565,
"learning_rate": 1.9634411306298614e-06,
"loss": 0.4981,
"step": 794
},
{
"epoch": 3.5990922844175492,
"grad_norm": 0.9187857508659363,
"learning_rate": 1.9571552451450542e-06,
"loss": 0.5318,
"step": 795
},
{
"epoch": 3.6036308623298035,
"grad_norm": 0.7069258689880371,
"learning_rate": 1.950872959270581e-06,
"loss": 0.4844,
"step": 796
},
{
"epoch": 3.608169440242057,
"grad_norm": 0.8754012584686279,
"learning_rate": 1.944594314664358e-06,
"loss": 0.4946,
"step": 797
},
{
"epoch": 3.6127080181543114,
"grad_norm": 0.8324469327926636,
"learning_rate": 1.938319352960156e-06,
"loss": 0.5172,
"step": 798
},
{
"epoch": 3.6172465960665656,
"grad_norm": 0.7333508729934692,
"learning_rate": 1.9320481157673225e-06,
"loss": 0.5157,
"step": 799
},
{
"epoch": 3.62178517397882,
"grad_norm": 0.7321659326553345,
"learning_rate": 1.9257806446705116e-06,
"loss": 0.5297,
"step": 800
},
{
"epoch": 3.626323751891074,
"grad_norm": 0.7912946343421936,
"learning_rate": 1.919516981229401e-06,
"loss": 0.5017,
"step": 801
},
{
"epoch": 3.6308623298033282,
"grad_norm": 0.7116424441337585,
"learning_rate": 1.9132571669784215e-06,
"loss": 0.502,
"step": 802
},
{
"epoch": 3.6354009077155824,
"grad_norm": 0.6774429678916931,
"learning_rate": 1.9070012434264793e-06,
"loss": 0.547,
"step": 803
},
{
"epoch": 3.6399394856278366,
"grad_norm": 0.9747133851051331,
"learning_rate": 1.9007492520566813e-06,
"loss": 0.512,
"step": 804
},
{
"epoch": 3.644478063540091,
"grad_norm": 0.6881934404373169,
"learning_rate": 1.8945012343260605e-06,
"loss": 0.5357,
"step": 805
},
{
"epoch": 3.649016641452345,
"grad_norm": 0.7221348881721497,
"learning_rate": 1.8882572316653003e-06,
"loss": 0.5118,
"step": 806
},
{
"epoch": 3.6535552193645993,
"grad_norm": 0.7595603466033936,
"learning_rate": 1.88201728547846e-06,
"loss": 0.5288,
"step": 807
},
{
"epoch": 3.6580937972768535,
"grad_norm": 0.9129202961921692,
"learning_rate": 1.8757814371427003e-06,
"loss": 0.4976,
"step": 808
},
{
"epoch": 3.6626323751891077,
"grad_norm": 1.1038625240325928,
"learning_rate": 1.8695497280080094e-06,
"loss": 0.5428,
"step": 809
},
{
"epoch": 3.6671709531013614,
"grad_norm": 0.9567497372627258,
"learning_rate": 1.8633221993969285e-06,
"loss": 0.498,
"step": 810
},
{
"epoch": 3.6717095310136156,
"grad_norm": 0.7735843062400818,
"learning_rate": 1.857098892604277e-06,
"loss": 0.5222,
"step": 811
},
{
"epoch": 3.67624810892587,
"grad_norm": 0.9261084794998169,
"learning_rate": 1.8508798488968805e-06,
"loss": 0.4969,
"step": 812
},
{
"epoch": 3.680786686838124,
"grad_norm": 0.7963117361068726,
"learning_rate": 1.844665109513294e-06,
"loss": 0.5127,
"step": 813
},
{
"epoch": 3.6853252647503782,
"grad_norm": 0.9042128920555115,
"learning_rate": 1.8384547156635324e-06,
"loss": 0.5214,
"step": 814
},
{
"epoch": 3.6898638426626325,
"grad_norm": 0.9819813370704651,
"learning_rate": 1.8322487085287953e-06,
"loss": 0.5208,
"step": 815
},
{
"epoch": 3.6944024205748867,
"grad_norm": 0.7022226452827454,
"learning_rate": 1.8260471292611936e-06,
"loss": 0.5422,
"step": 816
},
{
"epoch": 3.6989409984871404,
"grad_norm": 0.7549095749855042,
"learning_rate": 1.8198500189834757e-06,
"loss": 0.5197,
"step": 817
},
{
"epoch": 3.7034795763993946,
"grad_norm": 1.0044478178024292,
"learning_rate": 1.813657418788757e-06,
"loss": 0.5134,
"step": 818
},
{
"epoch": 3.708018154311649,
"grad_norm": 0.7384189963340759,
"learning_rate": 1.8074693697402473e-06,
"loss": 0.5367,
"step": 819
},
{
"epoch": 3.712556732223903,
"grad_norm": 0.7035711407661438,
"learning_rate": 1.8012859128709766e-06,
"loss": 0.509,
"step": 820
},
{
"epoch": 3.7170953101361572,
"grad_norm": 4.242894649505615,
"learning_rate": 1.7951070891835245e-06,
"loss": 0.4983,
"step": 821
},
{
"epoch": 3.7216338880484114,
"grad_norm": 0.7296944856643677,
"learning_rate": 1.7889329396497478e-06,
"loss": 0.5133,
"step": 822
},
{
"epoch": 3.7261724659606656,
"grad_norm": 0.6937296390533447,
"learning_rate": 1.7827635052105095e-06,
"loss": 0.5226,
"step": 823
},
{
"epoch": 3.73071104387292,
"grad_norm": 0.6841968894004822,
"learning_rate": 1.7765988267754053e-06,
"loss": 0.5287,
"step": 824
},
{
"epoch": 3.735249621785174,
"grad_norm": 0.7006022930145264,
"learning_rate": 1.7704389452224945e-06,
"loss": 0.5037,
"step": 825
},
{
"epoch": 3.7397881996974283,
"grad_norm": 1.5364707708358765,
"learning_rate": 1.7642839013980305e-06,
"loss": 0.4777,
"step": 826
},
{
"epoch": 3.7443267776096825,
"grad_norm": 0.7404047846794128,
"learning_rate": 1.7581337361161838e-06,
"loss": 0.5111,
"step": 827
},
{
"epoch": 3.7488653555219367,
"grad_norm": 0.6842325925827026,
"learning_rate": 1.7519884901587773e-06,
"loss": 0.5349,
"step": 828
},
{
"epoch": 3.753403933434191,
"grad_norm": 1.1200604438781738,
"learning_rate": 1.7458482042750138e-06,
"loss": 0.4894,
"step": 829
},
{
"epoch": 3.7579425113464446,
"grad_norm": 0.6718960404396057,
"learning_rate": 1.7397129191812058e-06,
"loss": 0.4482,
"step": 830
},
{
"epoch": 3.762481089258699,
"grad_norm": 0.7164394855499268,
"learning_rate": 1.7335826755605043e-06,
"loss": 0.5047,
"step": 831
},
{
"epoch": 3.767019667170953,
"grad_norm": 0.8456952571868896,
"learning_rate": 1.7274575140626318e-06,
"loss": 0.4874,
"step": 832
},
{
"epoch": 3.7715582450832073,
"grad_norm": 0.8413048386573792,
"learning_rate": 1.7213374753036105e-06,
"loss": 0.52,
"step": 833
},
{
"epoch": 3.7760968229954615,
"grad_norm": 0.7125583291053772,
"learning_rate": 1.7152225998654934e-06,
"loss": 0.4917,
"step": 834
},
{
"epoch": 3.7806354009077157,
"grad_norm": 0.7068613767623901,
"learning_rate": 1.7091129282960966e-06,
"loss": 0.5111,
"step": 835
},
{
"epoch": 3.78517397881997,
"grad_norm": 0.7550258636474609,
"learning_rate": 1.703008501108726e-06,
"loss": 0.5303,
"step": 836
},
{
"epoch": 3.789712556732224,
"grad_norm": 0.7415627241134644,
"learning_rate": 1.696909358781917e-06,
"loss": 0.5094,
"step": 837
},
{
"epoch": 3.794251134644478,
"grad_norm": 0.6911723017692566,
"learning_rate": 1.6908155417591576e-06,
"loss": 0.4979,
"step": 838
},
{
"epoch": 3.798789712556732,
"grad_norm": 0.7296070456504822,
"learning_rate": 1.684727090448624e-06,
"loss": 0.5099,
"step": 839
},
{
"epoch": 3.8033282904689862,
"grad_norm": 0.7332204580307007,
"learning_rate": 1.6786440452229134e-06,
"loss": 0.5198,
"step": 840
},
{
"epoch": 3.8078668683812404,
"grad_norm": 0.8778614401817322,
"learning_rate": 1.6725664464187734e-06,
"loss": 0.4959,
"step": 841
},
{
"epoch": 3.8124054462934946,
"grad_norm": 0.7139490246772766,
"learning_rate": 1.6664943343368386e-06,
"loss": 0.4968,
"step": 842
},
{
"epoch": 3.816944024205749,
"grad_norm": 0.7621096968650818,
"learning_rate": 1.660427749241359e-06,
"loss": 0.4932,
"step": 843
},
{
"epoch": 3.821482602118003,
"grad_norm": 0.6902374625205994,
"learning_rate": 1.6543667313599366e-06,
"loss": 0.5035,
"step": 844
},
{
"epoch": 3.8260211800302573,
"grad_norm": 0.6864549517631531,
"learning_rate": 1.6483113208832562e-06,
"loss": 0.5058,
"step": 845
},
{
"epoch": 3.8305597579425115,
"grad_norm": 0.7007944583892822,
"learning_rate": 1.6422615579648202e-06,
"loss": 0.5303,
"step": 846
},
{
"epoch": 3.8350983358547657,
"grad_norm": 5.041836738586426,
"learning_rate": 1.6362174827206806e-06,
"loss": 0.5142,
"step": 847
},
{
"epoch": 3.83963691376702,
"grad_norm": 0.7452664971351624,
"learning_rate": 1.6301791352291774e-06,
"loss": 0.4758,
"step": 848
},
{
"epoch": 3.844175491679274,
"grad_norm": 0.777967095375061,
"learning_rate": 1.6241465555306679e-06,
"loss": 0.5047,
"step": 849
},
{
"epoch": 3.848714069591528,
"grad_norm": 0.723528265953064,
"learning_rate": 1.618119783627263e-06,
"loss": 0.5263,
"step": 850
},
{
"epoch": 3.853252647503782,
"grad_norm": 0.6719141006469727,
"learning_rate": 1.612098859482562e-06,
"loss": 0.5024,
"step": 851
},
{
"epoch": 3.8577912254160363,
"grad_norm": 0.7341691255569458,
"learning_rate": 1.6060838230213883e-06,
"loss": 0.5148,
"step": 852
},
{
"epoch": 3.8623298033282905,
"grad_norm": 0.7549681663513184,
"learning_rate": 1.6000747141295233e-06,
"loss": 0.5314,
"step": 853
},
{
"epoch": 3.8668683812405447,
"grad_norm": 0.7185303568840027,
"learning_rate": 1.594071572653444e-06,
"loss": 0.5095,
"step": 854
},
{
"epoch": 3.871406959152799,
"grad_norm": 0.7704632878303528,
"learning_rate": 1.5880744384000544e-06,
"loss": 0.4827,
"step": 855
},
{
"epoch": 3.875945537065053,
"grad_norm": 0.7089985609054565,
"learning_rate": 1.5820833511364275e-06,
"loss": 0.5108,
"step": 856
},
{
"epoch": 3.8804841149773073,
"grad_norm": 0.7662360072135925,
"learning_rate": 1.5760983505895377e-06,
"loss": 0.5123,
"step": 857
},
{
"epoch": 3.885022692889561,
"grad_norm": 0.8514230847358704,
"learning_rate": 1.570119476445997e-06,
"loss": 0.5043,
"step": 858
},
{
"epoch": 3.8895612708018152,
"grad_norm": 0.6921412348747253,
"learning_rate": 1.5641467683517967e-06,
"loss": 0.4982,
"step": 859
},
{
"epoch": 3.8940998487140694,
"grad_norm": 0.6514143347740173,
"learning_rate": 1.558180265912037e-06,
"loss": 0.5285,
"step": 860
},
{
"epoch": 3.8986384266263236,
"grad_norm": 1.3229821920394897,
"learning_rate": 1.5522200086906708e-06,
"loss": 0.5135,
"step": 861
},
{
"epoch": 3.903177004538578,
"grad_norm": 0.6929940581321716,
"learning_rate": 1.5462660362102371e-06,
"loss": 0.5027,
"step": 862
},
{
"epoch": 3.907715582450832,
"grad_norm": 0.8205671906471252,
"learning_rate": 1.5403183879516025e-06,
"loss": 0.5031,
"step": 863
},
{
"epoch": 3.9122541603630863,
"grad_norm": 0.7823313474655151,
"learning_rate": 1.534377103353696e-06,
"loss": 0.5212,
"step": 864
},
{
"epoch": 3.9167927382753405,
"grad_norm": 1.095670223236084,
"learning_rate": 1.5284422218132495e-06,
"loss": 0.4762,
"step": 865
},
{
"epoch": 3.9213313161875947,
"grad_norm": 0.7315182685852051,
"learning_rate": 1.5225137826845371e-06,
"loss": 0.5012,
"step": 866
},
{
"epoch": 3.925869894099849,
"grad_norm": 1.2629145383834839,
"learning_rate": 1.5165918252791125e-06,
"loss": 0.488,
"step": 867
},
{
"epoch": 3.930408472012103,
"grad_norm": 0.8126862049102783,
"learning_rate": 1.510676388865548e-06,
"loss": 0.4999,
"step": 868
},
{
"epoch": 3.9349470499243573,
"grad_norm": 0.7050077319145203,
"learning_rate": 1.5047675126691783e-06,
"loss": 0.4863,
"step": 869
},
{
"epoch": 3.939485627836611,
"grad_norm": 0.8243216276168823,
"learning_rate": 1.4988652358718336e-06,
"loss": 0.5145,
"step": 870
},
{
"epoch": 3.9440242057488653,
"grad_norm": 0.9307572245597839,
"learning_rate": 1.4929695976115854e-06,
"loss": 0.5001,
"step": 871
},
{
"epoch": 3.9485627836611195,
"grad_norm": 0.6967872381210327,
"learning_rate": 1.4870806369824847e-06,
"loss": 0.5402,
"step": 872
},
{
"epoch": 3.9531013615733737,
"grad_norm": 0.7018783092498779,
"learning_rate": 1.4811983930343018e-06,
"loss": 0.5262,
"step": 873
},
{
"epoch": 3.957639939485628,
"grad_norm": 0.6808320879936218,
"learning_rate": 1.4753229047722704e-06,
"loss": 0.5046,
"step": 874
},
{
"epoch": 3.962178517397882,
"grad_norm": 1.0063635110855103,
"learning_rate": 1.4694542111568261e-06,
"loss": 0.5292,
"step": 875
},
{
"epoch": 3.9667170953101363,
"grad_norm": 0.7578749060630798,
"learning_rate": 1.4635923511033496e-06,
"loss": 0.4964,
"step": 876
},
{
"epoch": 3.9712556732223905,
"grad_norm": 0.7392986416816711,
"learning_rate": 1.4577373634819075e-06,
"loss": 0.5081,
"step": 877
},
{
"epoch": 3.9757942511346442,
"grad_norm": 0.737310528755188,
"learning_rate": 1.451889287116996e-06,
"loss": 0.5507,
"step": 878
},
{
"epoch": 3.9803328290468984,
"grad_norm": 1.102256417274475,
"learning_rate": 1.446048160787282e-06,
"loss": 0.5041,
"step": 879
},
{
"epoch": 3.9848714069591527,
"grad_norm": 0.790909469127655,
"learning_rate": 1.4402140232253486e-06,
"loss": 0.5064,
"step": 880
},
{
"epoch": 3.989409984871407,
"grad_norm": 0.6738175749778748,
"learning_rate": 1.4343869131174323e-06,
"loss": 0.4915,
"step": 881
},
{
"epoch": 3.993948562783661,
"grad_norm": 1.0193603038787842,
"learning_rate": 1.4285668691031751e-06,
"loss": 0.5251,
"step": 882
},
{
"epoch": 3.9984871406959153,
"grad_norm": 0.7585704922676086,
"learning_rate": 1.422753929775359e-06,
"loss": 0.5448,
"step": 883
},
{
"epoch": 4.0,
"grad_norm": 0.7585704922676086,
"learning_rate": 1.4169481336796598e-06,
"loss": 0.175,
"step": 884
},
{
"epoch": 4.004538577912254,
"grad_norm": 0.6884925365447998,
"learning_rate": 1.411149519314381e-06,
"loss": 0.4632,
"step": 885
},
{
"epoch": 4.009077155824508,
"grad_norm": 0.7509837746620178,
"learning_rate": 1.405358125130209e-06,
"loss": 0.54,
"step": 886
},
{
"epoch": 4.013615733736763,
"grad_norm": 0.6542647480964661,
"learning_rate": 1.399573989529949e-06,
"loss": 0.4786,
"step": 887
},
{
"epoch": 4.018154311649017,
"grad_norm": 0.8852865099906921,
"learning_rate": 1.393797150868278e-06,
"loss": 0.4871,
"step": 888
},
{
"epoch": 4.022692889561271,
"grad_norm": 0.6787682771682739,
"learning_rate": 1.3880276474514841e-06,
"loss": 0.5156,
"step": 889
},
{
"epoch": 4.027231467473525,
"grad_norm": 1.1594585180282593,
"learning_rate": 1.3822655175372148e-06,
"loss": 0.4806,
"step": 890
},
{
"epoch": 4.031770045385779,
"grad_norm": 1.0756573677062988,
"learning_rate": 1.3765107993342292e-06,
"loss": 0.4593,
"step": 891
},
{
"epoch": 4.036308623298034,
"grad_norm": 0.8092817068099976,
"learning_rate": 1.370763531002132e-06,
"loss": 0.5245,
"step": 892
},
{
"epoch": 4.040847201210288,
"grad_norm": 0.6932277679443359,
"learning_rate": 1.3650237506511333e-06,
"loss": 0.4957,
"step": 893
},
{
"epoch": 4.045385779122542,
"grad_norm": 0.7002460360527039,
"learning_rate": 1.3592914963417864e-06,
"loss": 0.4924,
"step": 894
},
{
"epoch": 4.049924357034795,
"grad_norm": 0.7570782899856567,
"learning_rate": 1.3535668060847428e-06,
"loss": 0.4914,
"step": 895
},
{
"epoch": 4.05446293494705,
"grad_norm": 0.8757203817367554,
"learning_rate": 1.347849717840493e-06,
"loss": 0.4971,
"step": 896
},
{
"epoch": 4.059001512859304,
"grad_norm": 0.7901989817619324,
"learning_rate": 1.3421402695191227e-06,
"loss": 0.5,
"step": 897
},
{
"epoch": 4.063540090771558,
"grad_norm": 0.7769845724105835,
"learning_rate": 1.3364384989800522e-06,
"loss": 0.4795,
"step": 898
},
{
"epoch": 4.068078668683812,
"grad_norm": 0.78565514087677,
"learning_rate": 1.3307444440317956e-06,
"loss": 0.4727,
"step": 899
},
{
"epoch": 4.072617246596066,
"grad_norm": 0.8624851703643799,
"learning_rate": 1.3250581424317012e-06,
"loss": 0.5184,
"step": 900
},
{
"epoch": 4.077155824508321,
"grad_norm": 0.7443495988845825,
"learning_rate": 1.3193796318857031e-06,
"loss": 0.474,
"step": 901
},
{
"epoch": 4.081694402420575,
"grad_norm": 0.8881354331970215,
"learning_rate": 1.3137089500480802e-06,
"loss": 0.4846,
"step": 902
},
{
"epoch": 4.086232980332829,
"grad_norm": 0.6768248081207275,
"learning_rate": 1.3080461345211909e-06,
"loss": 0.4898,
"step": 903
},
{
"epoch": 4.090771558245083,
"grad_norm": 0.7537845969200134,
"learning_rate": 1.3023912228552383e-06,
"loss": 0.4985,
"step": 904
},
{
"epoch": 4.095310136157337,
"grad_norm": 0.8789769411087036,
"learning_rate": 1.2967442525480092e-06,
"loss": 0.4818,
"step": 905
},
{
"epoch": 4.099848714069592,
"grad_norm": 0.8388983011245728,
"learning_rate": 1.2911052610446367e-06,
"loss": 0.495,
"step": 906
},
{
"epoch": 4.104387291981846,
"grad_norm": 0.7456345558166504,
"learning_rate": 1.2854742857373413e-06,
"loss": 0.4672,
"step": 907
},
{
"epoch": 4.1089258698941,
"grad_norm": 0.6598865985870361,
"learning_rate": 1.279851363965193e-06,
"loss": 0.467,
"step": 908
},
{
"epoch": 4.113464447806354,
"grad_norm": 0.7831140160560608,
"learning_rate": 1.2742365330138546e-06,
"loss": 0.4881,
"step": 909
},
{
"epoch": 4.118003025718608,
"grad_norm": 0.8532465696334839,
"learning_rate": 1.2686298301153394e-06,
"loss": 0.5227,
"step": 910
},
{
"epoch": 4.122541603630863,
"grad_norm": 0.8814959526062012,
"learning_rate": 1.2630312924477662e-06,
"loss": 0.4924,
"step": 911
},
{
"epoch": 4.127080181543117,
"grad_norm": 0.8461480140686035,
"learning_rate": 1.2574409571351048e-06,
"loss": 0.5074,
"step": 912
},
{
"epoch": 4.131618759455371,
"grad_norm": 0.8556418418884277,
"learning_rate": 1.2518588612469423e-06,
"loss": 0.4854,
"step": 913
},
{
"epoch": 4.136157337367624,
"grad_norm": 1.0551602840423584,
"learning_rate": 1.2462850417982234e-06,
"loss": 0.4865,
"step": 914
},
{
"epoch": 4.140695915279879,
"grad_norm": 1.2184538841247559,
"learning_rate": 1.2407195357490163e-06,
"loss": 0.4871,
"step": 915
},
{
"epoch": 4.145234493192133,
"grad_norm": 0.6897459030151367,
"learning_rate": 1.2351623800042586e-06,
"loss": 0.5006,
"step": 916
},
{
"epoch": 4.149773071104387,
"grad_norm": 0.8268042802810669,
"learning_rate": 1.2296136114135215e-06,
"loss": 0.4919,
"step": 917
},
{
"epoch": 4.154311649016641,
"grad_norm": 0.9411110281944275,
"learning_rate": 1.2240732667707581e-06,
"loss": 0.485,
"step": 918
},
{
"epoch": 4.158850226928895,
"grad_norm": 0.9568235278129578,
"learning_rate": 1.218541382814061e-06,
"loss": 0.4916,
"step": 919
},
{
"epoch": 4.16338880484115,
"grad_norm": 1.0960578918457031,
"learning_rate": 1.213017996225424e-06,
"loss": 0.5029,
"step": 920
},
{
"epoch": 4.167927382753404,
"grad_norm": 0.7468828558921814,
"learning_rate": 1.2075031436304906e-06,
"loss": 0.4882,
"step": 921
},
{
"epoch": 4.172465960665658,
"grad_norm": 0.9359505772590637,
"learning_rate": 1.2019968615983187e-06,
"loss": 0.4869,
"step": 922
},
{
"epoch": 4.177004538577912,
"grad_norm": 0.7222769260406494,
"learning_rate": 1.1964991866411304e-06,
"loss": 0.4573,
"step": 923
},
{
"epoch": 4.181543116490166,
"grad_norm": 0.7045331597328186,
"learning_rate": 1.1910101552140808e-06,
"loss": 0.5192,
"step": 924
},
{
"epoch": 4.186081694402421,
"grad_norm": 0.8232183456420898,
"learning_rate": 1.1855298037150022e-06,
"loss": 0.5022,
"step": 925
},
{
"epoch": 4.190620272314675,
"grad_norm": 0.7209495902061462,
"learning_rate": 1.1800581684841765e-06,
"loss": 0.4956,
"step": 926
},
{
"epoch": 4.195158850226929,
"grad_norm": 0.838749349117279,
"learning_rate": 1.1745952858040834e-06,
"loss": 0.4971,
"step": 927
},
{
"epoch": 4.199697428139183,
"grad_norm": 0.6986132264137268,
"learning_rate": 1.1691411918991657e-06,
"loss": 0.4914,
"step": 928
},
{
"epoch": 4.204236006051437,
"grad_norm": 0.7277268767356873,
"learning_rate": 1.1636959229355894e-06,
"loss": 0.4645,
"step": 929
},
{
"epoch": 4.208774583963692,
"grad_norm": 0.7252954244613647,
"learning_rate": 1.158259515020999e-06,
"loss": 0.5153,
"step": 930
},
{
"epoch": 4.213313161875946,
"grad_norm": 0.6610934734344482,
"learning_rate": 1.1528320042042853e-06,
"loss": 0.4828,
"step": 931
},
{
"epoch": 4.2178517397882,
"grad_norm": 1.002673864364624,
"learning_rate": 1.1474134264753384e-06,
"loss": 0.4962,
"step": 932
},
{
"epoch": 4.222390317700454,
"grad_norm": 0.7143538594245911,
"learning_rate": 1.142003817764816e-06,
"loss": 0.4798,
"step": 933
},
{
"epoch": 4.2269288956127085,
"grad_norm": 0.8639364242553711,
"learning_rate": 1.1366032139438987e-06,
"loss": 0.4641,
"step": 934
},
{
"epoch": 4.231467473524962,
"grad_norm": 0.8112154603004456,
"learning_rate": 1.1312116508240612e-06,
"loss": 0.4971,
"step": 935
},
{
"epoch": 4.236006051437216,
"grad_norm": 0.7791656851768494,
"learning_rate": 1.1258291641568237e-06,
"loss": 0.4963,
"step": 936
},
{
"epoch": 4.24054462934947,
"grad_norm": 0.849223256111145,
"learning_rate": 1.1204557896335217e-06,
"loss": 0.4826,
"step": 937
},
{
"epoch": 4.245083207261724,
"grad_norm": 0.6998844146728516,
"learning_rate": 1.1150915628850702e-06,
"loss": 0.4996,
"step": 938
},
{
"epoch": 4.249621785173979,
"grad_norm": 0.7256484627723694,
"learning_rate": 1.1097365194817222e-06,
"loss": 0.4219,
"step": 939
},
{
"epoch": 4.254160363086233,
"grad_norm": 0.7923110723495483,
"learning_rate": 1.1043906949328387e-06,
"loss": 0.4731,
"step": 940
},
{
"epoch": 4.258698940998487,
"grad_norm": 0.9490653872489929,
"learning_rate": 1.0990541246866473e-06,
"loss": 0.4884,
"step": 941
},
{
"epoch": 4.263237518910741,
"grad_norm": 1.0265283584594727,
"learning_rate": 1.0937268441300136e-06,
"loss": 0.4991,
"step": 942
},
{
"epoch": 4.267776096822995,
"grad_norm": 0.938533365726471,
"learning_rate": 1.088408888588199e-06,
"loss": 0.4748,
"step": 943
},
{
"epoch": 4.27231467473525,
"grad_norm": 0.6923869252204895,
"learning_rate": 1.083100293324636e-06,
"loss": 0.4721,
"step": 944
},
{
"epoch": 4.276853252647504,
"grad_norm": 0.7291610836982727,
"learning_rate": 1.0778010935406826e-06,
"loss": 0.5147,
"step": 945
},
{
"epoch": 4.281391830559758,
"grad_norm": 0.988376259803772,
"learning_rate": 1.0725113243754009e-06,
"loss": 0.4504,
"step": 946
},
{
"epoch": 4.285930408472012,
"grad_norm": 0.7914155125617981,
"learning_rate": 1.067231020905316e-06,
"loss": 0.4973,
"step": 947
},
{
"epoch": 4.290468986384266,
"grad_norm": 2.622145175933838,
"learning_rate": 1.061960218144185e-06,
"loss": 0.4732,
"step": 948
},
{
"epoch": 4.295007564296521,
"grad_norm": 0.722761332988739,
"learning_rate": 1.0566989510427678e-06,
"loss": 0.4759,
"step": 949
},
{
"epoch": 4.299546142208775,
"grad_norm": 0.7668985724449158,
"learning_rate": 1.051447254488591e-06,
"loss": 0.4969,
"step": 950
},
{
"epoch": 4.304084720121029,
"grad_norm": 0.9016123414039612,
"learning_rate": 1.0462051633057211e-06,
"loss": 0.4827,
"step": 951
},
{
"epoch": 4.308623298033283,
"grad_norm": 0.7125791907310486,
"learning_rate": 1.0409727122545285e-06,
"loss": 0.4691,
"step": 952
},
{
"epoch": 4.3131618759455375,
"grad_norm": 0.7095142006874084,
"learning_rate": 1.0357499360314632e-06,
"loss": 0.4705,
"step": 953
},
{
"epoch": 4.317700453857791,
"grad_norm": 1.2595562934875488,
"learning_rate": 1.0305368692688175e-06,
"loss": 0.4949,
"step": 954
},
{
"epoch": 4.322239031770045,
"grad_norm": 1.0172781944274902,
"learning_rate": 1.0253335465345037e-06,
"loss": 0.4448,
"step": 955
},
{
"epoch": 4.326777609682299,
"grad_norm": 0.8226476907730103,
"learning_rate": 1.0201400023318184e-06,
"loss": 0.5028,
"step": 956
},
{
"epoch": 4.331316187594553,
"grad_norm": 0.9284406304359436,
"learning_rate": 1.0149562710992184e-06,
"loss": 0.4971,
"step": 957
},
{
"epoch": 4.335854765506808,
"grad_norm": 2.741795063018799,
"learning_rate": 1.0097823872100913e-06,
"loss": 0.5247,
"step": 958
},
{
"epoch": 4.340393343419062,
"grad_norm": 0.8518880605697632,
"learning_rate": 1.0046183849725233e-06,
"loss": 0.5039,
"step": 959
},
{
"epoch": 4.344931921331316,
"grad_norm": 0.8044947981834412,
"learning_rate": 9.994642986290797e-07,
"loss": 0.4951,
"step": 960
},
{
"epoch": 4.34947049924357,
"grad_norm": 1.1144733428955078,
"learning_rate": 9.943201623565698e-07,
"loss": 0.4707,
"step": 961
},
{
"epoch": 4.354009077155824,
"grad_norm": 0.8300093412399292,
"learning_rate": 9.891860102658267e-07,
"loss": 0.4889,
"step": 962
},
{
"epoch": 4.358547655068079,
"grad_norm": 0.9095519781112671,
"learning_rate": 9.840618764014756e-07,
"loss": 0.4901,
"step": 963
},
{
"epoch": 4.363086232980333,
"grad_norm": 0.7411296963691711,
"learning_rate": 9.78947794741713e-07,
"loss": 0.4783,
"step": 964
},
{
"epoch": 4.367624810892587,
"grad_norm": 0.9816136360168457,
"learning_rate": 9.73843799198077e-07,
"loss": 0.4937,
"step": 965
},
{
"epoch": 4.372163388804841,
"grad_norm": 0.7506483793258667,
"learning_rate": 9.687499236152262e-07,
"loss": 0.4725,
"step": 966
},
{
"epoch": 4.376701966717095,
"grad_norm": 1.1032840013504028,
"learning_rate": 9.63666201770714e-07,
"loss": 0.4978,
"step": 967
},
{
"epoch": 4.38124054462935,
"grad_norm": 0.6765017509460449,
"learning_rate": 9.585926673747617e-07,
"loss": 0.4702,
"step": 968
},
{
"epoch": 4.385779122541604,
"grad_norm": 0.71392422914505,
"learning_rate": 9.535293540700408e-07,
"loss": 0.4616,
"step": 969
},
{
"epoch": 4.390317700453858,
"grad_norm": 1.6125102043151855,
"learning_rate": 9.48476295431443e-07,
"loss": 0.5081,
"step": 970
},
{
"epoch": 4.394856278366112,
"grad_norm": 0.7444176077842712,
"learning_rate": 9.434335249658649e-07,
"loss": 0.4614,
"step": 971
},
{
"epoch": 4.3993948562783665,
"grad_norm": 0.6820570230484009,
"learning_rate": 9.384010761119788e-07,
"loss": 0.4625,
"step": 972
},
{
"epoch": 4.403933434190621,
"grad_norm": 0.6871091723442078,
"learning_rate": 9.333789822400172e-07,
"loss": 0.5177,
"step": 973
},
{
"epoch": 4.408472012102875,
"grad_norm": 0.7985364198684692,
"learning_rate": 9.283672766515455e-07,
"loss": 0.5195,
"step": 974
},
{
"epoch": 4.413010590015128,
"grad_norm": 0.7615556716918945,
"learning_rate": 9.233659925792476e-07,
"loss": 0.4599,
"step": 975
},
{
"epoch": 4.417549167927382,
"grad_norm": 0.754677414894104,
"learning_rate": 9.183751631866992e-07,
"loss": 0.4978,
"step": 976
},
{
"epoch": 4.422087745839637,
"grad_norm": 0.7244731187820435,
"learning_rate": 9.133948215681524e-07,
"loss": 0.4411,
"step": 977
},
{
"epoch": 4.426626323751891,
"grad_norm": 0.778304398059845,
"learning_rate": 9.084250007483158e-07,
"loss": 0.5019,
"step": 978
},
{
"epoch": 4.431164901664145,
"grad_norm": 1.0169285535812378,
"learning_rate": 9.034657336821312e-07,
"loss": 0.4648,
"step": 979
},
{
"epoch": 4.435703479576399,
"grad_norm": 0.7210962772369385,
"learning_rate": 8.985170532545623e-07,
"loss": 0.508,
"step": 980
},
{
"epoch": 4.440242057488653,
"grad_norm": 0.7974186539649963,
"learning_rate": 8.935789922803689e-07,
"loss": 0.504,
"step": 981
},
{
"epoch": 4.444780635400908,
"grad_norm": 0.6913062930107117,
"learning_rate": 8.886515835038967e-07,
"loss": 0.4903,
"step": 982
},
{
"epoch": 4.449319213313162,
"grad_norm": 0.7884510159492493,
"learning_rate": 8.837348595988526e-07,
"loss": 0.4826,
"step": 983
},
{
"epoch": 4.453857791225416,
"grad_norm": 0.7987708449363708,
"learning_rate": 8.78828853168096e-07,
"loss": 0.4764,
"step": 984
},
{
"epoch": 4.45839636913767,
"grad_norm": 0.8689277172088623,
"learning_rate": 8.739335967434151e-07,
"loss": 0.4681,
"step": 985
},
{
"epoch": 4.462934947049924,
"grad_norm": 0.8613301515579224,
"learning_rate": 8.690491227853177e-07,
"loss": 0.4696,
"step": 986
},
{
"epoch": 4.467473524962179,
"grad_norm": 1.1412184238433838,
"learning_rate": 8.6417546368281e-07,
"loss": 0.5196,
"step": 987
},
{
"epoch": 4.472012102874433,
"grad_norm": 0.7207121253013611,
"learning_rate": 8.593126517531869e-07,
"loss": 0.4838,
"step": 988
},
{
"epoch": 4.476550680786687,
"grad_norm": 3.387355327606201,
"learning_rate": 8.544607192418161e-07,
"loss": 0.4824,
"step": 989
},
{
"epoch": 4.481089258698941,
"grad_norm": 0.9257369637489319,
"learning_rate": 8.496196983219205e-07,
"loss": 0.4825,
"step": 990
},
{
"epoch": 4.4856278366111955,
"grad_norm": 0.8000094294548035,
"learning_rate": 8.447896210943718e-07,
"loss": 0.5123,
"step": 991
},
{
"epoch": 4.49016641452345,
"grad_norm": 0.6996716260910034,
"learning_rate": 8.399705195874708e-07,
"loss": 0.5056,
"step": 992
},
{
"epoch": 4.494704992435704,
"grad_norm": 0.688705563545227,
"learning_rate": 8.351624257567415e-07,
"loss": 0.4915,
"step": 993
},
{
"epoch": 4.499243570347957,
"grad_norm": 1.2094013690948486,
"learning_rate": 8.303653714847118e-07,
"loss": 0.4977,
"step": 994
},
{
"epoch": 4.503782148260212,
"grad_norm": 0.7317572832107544,
"learning_rate": 8.255793885807104e-07,
"loss": 0.4863,
"step": 995
},
{
"epoch": 4.508320726172466,
"grad_norm": 0.7524070739746094,
"learning_rate": 8.208045087806479e-07,
"loss": 0.5072,
"step": 996
},
{
"epoch": 4.51285930408472,
"grad_norm": 0.7206716537475586,
"learning_rate": 8.160407637468134e-07,
"loss": 0.4787,
"step": 997
},
{
"epoch": 4.517397881996974,
"grad_norm": 0.9675849080085754,
"learning_rate": 8.112881850676585e-07,
"loss": 0.5062,
"step": 998
},
{
"epoch": 4.521936459909228,
"grad_norm": 0.6949933767318726,
"learning_rate": 8.065468042575905e-07,
"loss": 0.508,
"step": 999
},
{
"epoch": 4.526475037821482,
"grad_norm": 0.9218463897705078,
"learning_rate": 8.018166527567672e-07,
"loss": 0.486,
"step": 1000
},
{
"epoch": 4.531013615733737,
"grad_norm": 0.7000327110290527,
"learning_rate": 7.970977619308806e-07,
"loss": 0.46,
"step": 1001
},
{
"epoch": 4.535552193645991,
"grad_norm": 0.740325391292572,
"learning_rate": 7.923901630709554e-07,
"loss": 0.4896,
"step": 1002
},
{
"epoch": 4.540090771558245,
"grad_norm": 1.6052324771881104,
"learning_rate": 7.87693887393137e-07,
"loss": 0.4878,
"step": 1003
},
{
"epoch": 4.544629349470499,
"grad_norm": 0.7649421095848083,
"learning_rate": 7.830089660384896e-07,
"loss": 0.4924,
"step": 1004
},
{
"epoch": 4.549167927382753,
"grad_norm": 0.734867513179779,
"learning_rate": 7.783354300727835e-07,
"loss": 0.4788,
"step": 1005
},
{
"epoch": 4.553706505295008,
"grad_norm": 0.6974482536315918,
"learning_rate": 7.736733104862953e-07,
"loss": 0.48,
"step": 1006
},
{
"epoch": 4.558245083207262,
"grad_norm": 0.7161325812339783,
"learning_rate": 7.690226381935976e-07,
"loss": 0.5301,
"step": 1007
},
{
"epoch": 4.562783661119516,
"grad_norm": 0.676956057548523,
"learning_rate": 7.643834440333553e-07,
"loss": 0.4803,
"step": 1008
},
{
"epoch": 4.56732223903177,
"grad_norm": 0.7367483973503113,
"learning_rate": 7.597557587681242e-07,
"loss": 0.4826,
"step": 1009
},
{
"epoch": 4.5718608169440245,
"grad_norm": 0.7037650942802429,
"learning_rate": 7.551396130841406e-07,
"loss": 0.4926,
"step": 1010
},
{
"epoch": 4.576399394856279,
"grad_norm": 0.6749675869941711,
"learning_rate": 7.505350375911278e-07,
"loss": 0.5091,
"step": 1011
},
{
"epoch": 4.580937972768533,
"grad_norm": 0.6664526462554932,
"learning_rate": 7.459420628220801e-07,
"loss": 0.4723,
"step": 1012
},
{
"epoch": 4.585476550680786,
"grad_norm": 0.7849709391593933,
"learning_rate": 7.413607192330724e-07,
"loss": 0.5035,
"step": 1013
},
{
"epoch": 4.590015128593041,
"grad_norm": 0.881906270980835,
"learning_rate": 7.367910372030495e-07,
"loss": 0.4922,
"step": 1014
},
{
"epoch": 4.594553706505295,
"grad_norm": 1.0520840883255005,
"learning_rate": 7.322330470336314e-07,
"loss": 0.4872,
"step": 1015
},
{
"epoch": 4.599092284417549,
"grad_norm": 0.7468703985214233,
"learning_rate": 7.27686778948907e-07,
"loss": 0.4962,
"step": 1016
},
{
"epoch": 4.603630862329803,
"grad_norm": 1.159132480621338,
"learning_rate": 7.231522630952359e-07,
"loss": 0.4881,
"step": 1017
},
{
"epoch": 4.608169440242057,
"grad_norm": 0.7285956144332886,
"learning_rate": 7.186295295410506e-07,
"loss": 0.5156,
"step": 1018
},
{
"epoch": 4.612708018154311,
"grad_norm": 0.7299140095710754,
"learning_rate": 7.141186082766521e-07,
"loss": 0.5128,
"step": 1019
},
{
"epoch": 4.617246596066566,
"grad_norm": 0.8349388837814331,
"learning_rate": 7.096195292140173e-07,
"loss": 0.4753,
"step": 1020
},
{
"epoch": 4.62178517397882,
"grad_norm": 0.7903823256492615,
"learning_rate": 7.051323221865933e-07,
"loss": 0.4785,
"step": 1021
},
{
"epoch": 4.626323751891074,
"grad_norm": 0.9033539295196533,
"learning_rate": 7.006570169491084e-07,
"loss": 0.517,
"step": 1022
},
{
"epoch": 4.630862329803328,
"grad_norm": 0.7931293845176697,
"learning_rate": 6.961936431773655e-07,
"loss": 0.4976,
"step": 1023
},
{
"epoch": 4.635400907715582,
"grad_norm": 2.131030559539795,
"learning_rate": 6.917422304680532e-07,
"loss": 0.4781,
"step": 1024
},
{
"epoch": 4.639939485627837,
"grad_norm": 0.881932258605957,
"learning_rate": 6.873028083385436e-07,
"loss": 0.4841,
"step": 1025
},
{
"epoch": 4.644478063540091,
"grad_norm": 0.7999169230461121,
"learning_rate": 6.828754062266996e-07,
"loss": 0.5095,
"step": 1026
},
{
"epoch": 4.649016641452345,
"grad_norm": 0.7111617922782898,
"learning_rate": 6.784600534906813e-07,
"loss": 0.4742,
"step": 1027
},
{
"epoch": 4.653555219364599,
"grad_norm": 0.7100422978401184,
"learning_rate": 6.740567794087463e-07,
"loss": 0.5097,
"step": 1028
},
{
"epoch": 4.6580937972768535,
"grad_norm": 0.7377627491950989,
"learning_rate": 6.696656131790611e-07,
"loss": 0.515,
"step": 1029
},
{
"epoch": 4.662632375189108,
"grad_norm": 0.7187225222587585,
"learning_rate": 6.652865839195025e-07,
"loss": 0.4902,
"step": 1030
},
{
"epoch": 4.667170953101362,
"grad_norm": 0.7497810125350952,
"learning_rate": 6.609197206674694e-07,
"loss": 0.4835,
"step": 1031
},
{
"epoch": 4.671709531013616,
"grad_norm": 0.7579675912857056,
"learning_rate": 6.565650523796843e-07,
"loss": 0.4518,
"step": 1032
},
{
"epoch": 4.67624810892587,
"grad_norm": 0.7889072299003601,
"learning_rate": 6.522226079320099e-07,
"loss": 0.4745,
"step": 1033
},
{
"epoch": 4.680786686838124,
"grad_norm": 0.8384209275245667,
"learning_rate": 6.478924161192479e-07,
"loss": 0.4639,
"step": 1034
},
{
"epoch": 4.685325264750379,
"grad_norm": 0.876602292060852,
"learning_rate": 6.435745056549533e-07,
"loss": 0.5324,
"step": 1035
},
{
"epoch": 4.689863842662632,
"grad_norm": 1.0869137048721313,
"learning_rate": 6.392689051712458e-07,
"loss": 0.4907,
"step": 1036
},
{
"epoch": 4.694402420574886,
"grad_norm": 0.7678726315498352,
"learning_rate": 6.349756432186143e-07,
"loss": 0.5281,
"step": 1037
},
{
"epoch": 4.69894099848714,
"grad_norm": 0.7951900959014893,
"learning_rate": 6.306947482657341e-07,
"loss": 0.4939,
"step": 1038
},
{
"epoch": 4.703479576399395,
"grad_norm": 0.7335869073867798,
"learning_rate": 6.264262486992712e-07,
"loss": 0.441,
"step": 1039
},
{
"epoch": 4.708018154311649,
"grad_norm": 0.6744561195373535,
"learning_rate": 6.221701728237008e-07,
"loss": 0.4914,
"step": 1040
},
{
"epoch": 4.712556732223903,
"grad_norm": 0.8650886416435242,
"learning_rate": 6.179265488611138e-07,
"loss": 0.5063,
"step": 1041
},
{
"epoch": 4.717095310136157,
"grad_norm": 0.7943755984306335,
"learning_rate": 6.136954049510351e-07,
"loss": 0.4598,
"step": 1042
},
{
"epoch": 4.721633888048411,
"grad_norm": 0.870184063911438,
"learning_rate": 6.09476769150231e-07,
"loss": 0.5063,
"step": 1043
},
{
"epoch": 4.726172465960666,
"grad_norm": 0.7460466027259827,
"learning_rate": 6.052706694325289e-07,
"loss": 0.5014,
"step": 1044
},
{
"epoch": 4.73071104387292,
"grad_norm": 1.1483376026153564,
"learning_rate": 6.010771336886292e-07,
"loss": 0.5028,
"step": 1045
},
{
"epoch": 4.735249621785174,
"grad_norm": 0.7843842506408691,
"learning_rate": 5.968961897259185e-07,
"loss": 0.4962,
"step": 1046
},
{
"epoch": 4.739788199697428,
"grad_norm": 0.7222856283187866,
"learning_rate": 5.927278652682891e-07,
"loss": 0.4836,
"step": 1047
},
{
"epoch": 4.7443267776096825,
"grad_norm": 0.969645619392395,
"learning_rate": 5.885721879559514e-07,
"loss": 0.4867,
"step": 1048
},
{
"epoch": 4.748865355521937,
"grad_norm": 1.05107581615448,
"learning_rate": 5.844291853452549e-07,
"loss": 0.4977,
"step": 1049
},
{
"epoch": 4.753403933434191,
"grad_norm": 0.6660315990447998,
"learning_rate": 5.802988849085001e-07,
"loss": 0.5086,
"step": 1050
},
{
"epoch": 4.757942511346445,
"grad_norm": 1.050270438194275,
"learning_rate": 5.761813140337619e-07,
"loss": 0.4938,
"step": 1051
},
{
"epoch": 4.762481089258699,
"grad_norm": 1.5800691843032837,
"learning_rate": 5.720765000247027e-07,
"loss": 0.5146,
"step": 1052
},
{
"epoch": 4.7670196671709535,
"grad_norm": 0.6957308650016785,
"learning_rate": 5.67984470100397e-07,
"loss": 0.4802,
"step": 1053
},
{
"epoch": 4.771558245083208,
"grad_norm": 13.844222068786621,
"learning_rate": 5.639052513951449e-07,
"loss": 0.5171,
"step": 1054
},
{
"epoch": 4.776096822995461,
"grad_norm": 1.097083330154419,
"learning_rate": 5.598388709582963e-07,
"loss": 0.4915,
"step": 1055
},
{
"epoch": 4.780635400907715,
"grad_norm": 0.8402828574180603,
"learning_rate": 5.557853557540724e-07,
"loss": 0.4767,
"step": 1056
},
{
"epoch": 4.785173978819969,
"grad_norm": 1.0703009366989136,
"learning_rate": 5.517447326613809e-07,
"loss": 0.4577,
"step": 1057
},
{
"epoch": 4.789712556732224,
"grad_norm": 0.824536919593811,
"learning_rate": 5.477170284736455e-07,
"loss": 0.4635,
"step": 1058
},
{
"epoch": 4.794251134644478,
"grad_norm": 0.7436249256134033,
"learning_rate": 5.437022698986217e-07,
"loss": 0.4882,
"step": 1059
},
{
"epoch": 4.798789712556732,
"grad_norm": 0.7780619859695435,
"learning_rate": 5.397004835582242e-07,
"loss": 0.4823,
"step": 1060
},
{
"epoch": 4.803328290468986,
"grad_norm": 0.9229158163070679,
"learning_rate": 5.357116959883477e-07,
"loss": 0.4963,
"step": 1061
},
{
"epoch": 4.80786686838124,
"grad_norm": 0.7092882990837097,
"learning_rate": 5.317359336386932e-07,
"loss": 0.5061,
"step": 1062
},
{
"epoch": 4.812405446293495,
"grad_norm": 0.8238071799278259,
"learning_rate": 5.27773222872589e-07,
"loss": 0.4716,
"step": 1063
},
{
"epoch": 4.816944024205749,
"grad_norm": 0.9120982885360718,
"learning_rate": 5.238235899668212e-07,
"loss": 0.4613,
"step": 1064
},
{
"epoch": 4.821482602118003,
"grad_norm": 0.8099984526634216,
"learning_rate": 5.198870611114529e-07,
"loss": 0.4949,
"step": 1065
},
{
"epoch": 4.826021180030257,
"grad_norm": 0.885421097278595,
"learning_rate": 5.159636624096572e-07,
"loss": 0.4697,
"step": 1066
},
{
"epoch": 4.8305597579425115,
"grad_norm": 0.8806759715080261,
"learning_rate": 5.120534198775404e-07,
"loss": 0.5057,
"step": 1067
},
{
"epoch": 4.835098335854766,
"grad_norm": 1.269351840019226,
"learning_rate": 5.081563594439676e-07,
"loss": 0.5088,
"step": 1068
},
{
"epoch": 4.83963691376702,
"grad_norm": 0.7071214318275452,
"learning_rate": 5.04272506950397e-07,
"loss": 0.4803,
"step": 1069
},
{
"epoch": 4.844175491679274,
"grad_norm": 0.762505054473877,
"learning_rate": 5.004018881507016e-07,
"loss": 0.4945,
"step": 1070
},
{
"epoch": 4.848714069591528,
"grad_norm": 0.792746365070343,
"learning_rate": 4.965445287110038e-07,
"loss": 0.5061,
"step": 1071
},
{
"epoch": 4.8532526475037825,
"grad_norm": 0.7810184955596924,
"learning_rate": 4.92700454209501e-07,
"loss": 0.4824,
"step": 1072
},
{
"epoch": 4.857791225416037,
"grad_norm": 0.7621875405311584,
"learning_rate": 4.888696901363005e-07,
"loss": 0.4603,
"step": 1073
},
{
"epoch": 4.86232980332829,
"grad_norm": 1.3374532461166382,
"learning_rate": 4.850522618932449e-07,
"loss": 0.5073,
"step": 1074
},
{
"epoch": 4.866868381240545,
"grad_norm": 0.7166194915771484,
"learning_rate": 4.812481947937498e-07,
"loss": 0.484,
"step": 1075
},
{
"epoch": 4.871406959152798,
"grad_norm": 1.4800307750701904,
"learning_rate": 4.774575140626317e-07,
"loss": 0.4728,
"step": 1076
},
{
"epoch": 4.875945537065053,
"grad_norm": 0.7257369160652161,
"learning_rate": 4.736802448359404e-07,
"loss": 0.4827,
"step": 1077
},
{
"epoch": 4.880484114977307,
"grad_norm": 0.6776279211044312,
"learning_rate": 4.6991641216079684e-07,
"loss": 0.5066,
"step": 1078
},
{
"epoch": 4.885022692889561,
"grad_norm": 0.7334598898887634,
"learning_rate": 4.6616604099522097e-07,
"loss": 0.4857,
"step": 1079
},
{
"epoch": 4.889561270801815,
"grad_norm": 0.6790256500244141,
"learning_rate": 4.624291562079719e-07,
"loss": 0.4921,
"step": 1080
},
{
"epoch": 4.8940998487140694,
"grad_norm": 0.7464075088500977,
"learning_rate": 4.5870578257837803e-07,
"loss": 0.4441,
"step": 1081
},
{
"epoch": 4.898638426626324,
"grad_norm": 0.7377055287361145,
"learning_rate": 4.549959447961777e-07,
"loss": 0.4607,
"step": 1082
},
{
"epoch": 4.903177004538578,
"grad_norm": 0.8656676411628723,
"learning_rate": 4.5129966746134987e-07,
"loss": 0.4524,
"step": 1083
},
{
"epoch": 4.907715582450832,
"grad_norm": 0.9937976002693176,
"learning_rate": 4.476169750839571e-07,
"loss": 0.4831,
"step": 1084
},
{
"epoch": 4.912254160363086,
"grad_norm": 0.8174405097961426,
"learning_rate": 4.439478920839771e-07,
"loss": 0.4755,
"step": 1085
},
{
"epoch": 4.9167927382753405,
"grad_norm": 0.7877658009529114,
"learning_rate": 4.402924427911459e-07,
"loss": 0.5249,
"step": 1086
},
{
"epoch": 4.921331316187595,
"grad_norm": 0.7509103417396545,
"learning_rate": 4.366506514447932e-07,
"loss": 0.5099,
"step": 1087
},
{
"epoch": 4.925869894099849,
"grad_norm": 0.7400006651878357,
"learning_rate": 4.3302254219368235e-07,
"loss": 0.4449,
"step": 1088
},
{
"epoch": 4.930408472012103,
"grad_norm": 0.7745682001113892,
"learning_rate": 4.294081390958521e-07,
"loss": 0.4939,
"step": 1089
},
{
"epoch": 4.934947049924357,
"grad_norm": 0.7397611141204834,
"learning_rate": 4.2580746611845273e-07,
"loss": 0.5219,
"step": 1090
},
{
"epoch": 4.9394856278366115,
"grad_norm": 0.8154904842376709,
"learning_rate": 4.2222054713759315e-07,
"loss": 0.4853,
"step": 1091
},
{
"epoch": 4.944024205748866,
"grad_norm": 0.8152079582214355,
"learning_rate": 4.186474059381768e-07,
"loss": 0.5134,
"step": 1092
},
{
"epoch": 4.94856278366112,
"grad_norm": 0.776576578617096,
"learning_rate": 4.1508806621374825e-07,
"loss": 0.4983,
"step": 1093
},
{
"epoch": 4.953101361573374,
"grad_norm": 0.729992151260376,
"learning_rate": 4.1154255156633267e-07,
"loss": 0.4978,
"step": 1094
},
{
"epoch": 4.957639939485627,
"grad_norm": 0.7563814520835876,
"learning_rate": 4.0801088550628307e-07,
"loss": 0.4817,
"step": 1095
},
{
"epoch": 4.962178517397882,
"grad_norm": 1.2037073373794556,
"learning_rate": 4.044930914521203e-07,
"loss": 0.455,
"step": 1096
},
{
"epoch": 4.966717095310136,
"grad_norm": 0.7728638648986816,
"learning_rate": 4.0098919273037965e-07,
"loss": 0.4898,
"step": 1097
},
{
"epoch": 4.97125567322239,
"grad_norm": 0.7198495268821716,
"learning_rate": 3.9749921257545926e-07,
"loss": 0.5035,
"step": 1098
},
{
"epoch": 4.975794251134644,
"grad_norm": 0.864120602607727,
"learning_rate": 3.940231741294598e-07,
"loss": 0.506,
"step": 1099
},
{
"epoch": 4.9803328290468984,
"grad_norm": 0.7351332306861877,
"learning_rate": 3.9056110044203594e-07,
"loss": 0.5028,
"step": 1100
},
{
"epoch": 4.984871406959153,
"grad_norm": 0.7667080163955688,
"learning_rate": 3.8711301447024093e-07,
"loss": 0.4922,
"step": 1101
},
{
"epoch": 4.989409984871407,
"grad_norm": 0.8356844186782837,
"learning_rate": 3.8367893907837683e-07,
"loss": 0.5058,
"step": 1102
},
{
"epoch": 4.993948562783661,
"grad_norm": 0.7784990072250366,
"learning_rate": 3.802588970378387e-07,
"loss": 0.4222,
"step": 1103
},
{
"epoch": 4.998487140695915,
"grad_norm": 0.7764464616775513,
"learning_rate": 3.7685291102696976e-07,
"loss": 0.4688,
"step": 1104
},
{
"epoch": 5.0,
"grad_norm": 0.7764464616775513,
"learning_rate": 3.734610036309047e-07,
"loss": 0.1767,
"step": 1105
},
{
"epoch": 5.004538577912254,
"grad_norm": 1.046483039855957,
"learning_rate": 3.7008319734142346e-07,
"loss": 0.4978,
"step": 1106
},
{
"epoch": 5.009077155824508,
"grad_norm": 0.7094864845275879,
"learning_rate": 3.6671951455680254e-07,
"loss": 0.45,
"step": 1107
},
{
"epoch": 5.013615733736763,
"grad_norm": 0.8907707333564758,
"learning_rate": 3.6336997758166263e-07,
"loss": 0.4951,
"step": 1108
},
{
"epoch": 5.018154311649017,
"grad_norm": 0.7162082195281982,
"learning_rate": 3.6003460862682774e-07,
"loss": 0.4909,
"step": 1109
},
{
"epoch": 5.022692889561271,
"grad_norm": 0.9485985636711121,
"learning_rate": 3.56713429809169e-07,
"loss": 0.4624,
"step": 1110
},
{
"epoch": 5.027231467473525,
"grad_norm": 0.7002444863319397,
"learning_rate": 3.534064631514658e-07,
"loss": 0.4572,
"step": 1111
},
{
"epoch": 5.031770045385779,
"grad_norm": 0.6732192635536194,
"learning_rate": 3.5011373058225413e-07,
"loss": 0.4476,
"step": 1112
},
{
"epoch": 5.036308623298034,
"grad_norm": 0.7230054140090942,
"learning_rate": 3.468352539356856e-07,
"loss": 0.4928,
"step": 1113
},
{
"epoch": 5.040847201210288,
"grad_norm": 0.8661540746688843,
"learning_rate": 3.435710549513793e-07,
"loss": 0.4815,
"step": 1114
},
{
"epoch": 5.045385779122542,
"grad_norm": 0.7192421555519104,
"learning_rate": 3.403211552742788e-07,
"loss": 0.4345,
"step": 1115
},
{
"epoch": 5.049924357034795,
"grad_norm": 0.775611162185669,
"learning_rate": 3.370855764545106e-07,
"loss": 0.4514,
"step": 1116
},
{
"epoch": 5.05446293494705,
"grad_norm": 0.666876494884491,
"learning_rate": 3.3386433994723717e-07,
"loss": 0.4622,
"step": 1117
},
{
"epoch": 5.059001512859304,
"grad_norm": 0.9268916249275208,
"learning_rate": 3.306574671125185e-07,
"loss": 0.4624,
"step": 1118
},
{
"epoch": 5.063540090771558,
"grad_norm": 1.5524824857711792,
"learning_rate": 3.274649792151671e-07,
"loss": 0.4963,
"step": 1119
},
{
"epoch": 5.068078668683812,
"grad_norm": 0.9165545701980591,
"learning_rate": 3.2428689742461187e-07,
"loss": 0.4685,
"step": 1120
},
{
"epoch": 5.072617246596066,
"grad_norm": 0.8844300508499146,
"learning_rate": 3.2112324281475174e-07,
"loss": 0.4791,
"step": 1121
},
{
"epoch": 5.077155824508321,
"grad_norm": 0.6754626631736755,
"learning_rate": 3.1797403636382035e-07,
"loss": 0.4844,
"step": 1122
},
{
"epoch": 5.081694402420575,
"grad_norm": 0.9919183850288391,
"learning_rate": 3.1483929895424453e-07,
"loss": 0.4755,
"step": 1123
},
{
"epoch": 5.086232980332829,
"grad_norm": 0.7306275963783264,
"learning_rate": 3.117190513725066e-07,
"loss": 0.4958,
"step": 1124
},
{
"epoch": 5.090771558245083,
"grad_norm": 0.7538716793060303,
"learning_rate": 3.086133143090081e-07,
"loss": 0.4927,
"step": 1125
},
{
"epoch": 5.095310136157337,
"grad_norm": 0.851882815361023,
"learning_rate": 3.055221083579296e-07,
"loss": 0.4807,
"step": 1126
},
{
"epoch": 5.099848714069592,
"grad_norm": 0.7952591776847839,
"learning_rate": 3.024454540170968e-07,
"loss": 0.4903,
"step": 1127
},
{
"epoch": 5.104387291981846,
"grad_norm": 0.7629356384277344,
"learning_rate": 2.993833716878422e-07,
"loss": 0.4954,
"step": 1128
},
{
"epoch": 5.1089258698941,
"grad_norm": 0.9297985434532166,
"learning_rate": 2.9633588167487266e-07,
"loss": 0.513,
"step": 1129
},
{
"epoch": 5.113464447806354,
"grad_norm": 0.6958643198013306,
"learning_rate": 2.933030041861312e-07,
"loss": 0.4906,
"step": 1130
},
{
"epoch": 5.118003025718608,
"grad_norm": 0.6689577698707581,
"learning_rate": 2.9028475933266825e-07,
"loss": 0.429,
"step": 1131
},
{
"epoch": 5.122541603630863,
"grad_norm": 0.8975169658660889,
"learning_rate": 2.8728116712850195e-07,
"loss": 0.464,
"step": 1132
},
{
"epoch": 5.127080181543117,
"grad_norm": 0.8286609649658203,
"learning_rate": 2.8429224749048884e-07,
"loss": 0.4497,
"step": 1133
},
{
"epoch": 5.131618759455371,
"grad_norm": 0.8133575320243835,
"learning_rate": 2.8131802023819376e-07,
"loss": 0.4943,
"step": 1134
},
{
"epoch": 5.136157337367624,
"grad_norm": 0.7451333403587341,
"learning_rate": 2.783585050937537e-07,
"loss": 0.4699,
"step": 1135
},
{
"epoch": 5.140695915279879,
"grad_norm": 0.9283474087715149,
"learning_rate": 2.7541372168175093e-07,
"loss": 0.4896,
"step": 1136
},
{
"epoch": 5.145234493192133,
"grad_norm": 0.7920743823051453,
"learning_rate": 2.7248368952908055e-07,
"loss": 0.4999,
"step": 1137
},
{
"epoch": 5.149773071104387,
"grad_norm": 0.7597651481628418,
"learning_rate": 2.695684280648228e-07,
"loss": 0.486,
"step": 1138
},
{
"epoch": 5.154311649016641,
"grad_norm": 0.7768572568893433,
"learning_rate": 2.666679566201119e-07,
"loss": 0.48,
"step": 1139
},
{
"epoch": 5.158850226928895,
"grad_norm": 0.9812193512916565,
"learning_rate": 2.6378229442801163e-07,
"loss": 0.4836,
"step": 1140
},
{
"epoch": 5.16338880484115,
"grad_norm": 0.7561526894569397,
"learning_rate": 2.6091146062338176e-07,
"loss": 0.4618,
"step": 1141
},
{
"epoch": 5.167927382753404,
"grad_norm": 0.711685299873352,
"learning_rate": 2.580554742427591e-07,
"loss": 0.4736,
"step": 1142
},
{
"epoch": 5.172465960665658,
"grad_norm": 0.712313711643219,
"learning_rate": 2.5521435422422443e-07,
"loss": 0.4547,
"step": 1143
},
{
"epoch": 5.177004538577912,
"grad_norm": 0.839168131351471,
"learning_rate": 2.523881194072791e-07,
"loss": 0.5002,
"step": 1144
},
{
"epoch": 5.181543116490166,
"grad_norm": 0.8970273733139038,
"learning_rate": 2.4957678853272246e-07,
"loss": 0.5014,
"step": 1145
},
{
"epoch": 5.186081694402421,
"grad_norm": 0.6730303764343262,
"learning_rate": 2.4678038024252363e-07,
"loss": 0.4691,
"step": 1146
},
{
"epoch": 5.190620272314675,
"grad_norm": 0.915203869342804,
"learning_rate": 2.4399891307970187e-07,
"loss": 0.4886,
"step": 1147
},
{
"epoch": 5.195158850226929,
"grad_norm": 0.7596525549888611,
"learning_rate": 2.4123240548819955e-07,
"loss": 0.474,
"step": 1148
},
{
"epoch": 5.199697428139183,
"grad_norm": 1.1874265670776367,
"learning_rate": 2.384808758127638e-07,
"loss": 0.4834,
"step": 1149
},
{
"epoch": 5.204236006051437,
"grad_norm": 0.6951465606689453,
"learning_rate": 2.357443422988215e-07,
"loss": 0.4736,
"step": 1150
},
{
"epoch": 5.208774583963692,
"grad_norm": 0.9026170969009399,
"learning_rate": 2.330228230923612e-07,
"loss": 0.4752,
"step": 1151
},
{
"epoch": 5.213313161875946,
"grad_norm": 0.9337056875228882,
"learning_rate": 2.3031633623980975e-07,
"loss": 0.4745,
"step": 1152
},
{
"epoch": 5.2178517397882,
"grad_norm": 0.6569899916648865,
"learning_rate": 2.2762489968791601e-07,
"loss": 0.4856,
"step": 1153
},
{
"epoch": 5.222390317700454,
"grad_norm": 0.7478249669075012,
"learning_rate": 2.249485312836294e-07,
"loss": 0.5025,
"step": 1154
},
{
"epoch": 5.2269288956127085,
"grad_norm": 0.7351001501083374,
"learning_rate": 2.2228724877398134e-07,
"loss": 0.4784,
"step": 1155
},
{
"epoch": 5.231467473524962,
"grad_norm": 0.7316746115684509,
"learning_rate": 2.1964106980597034e-07,
"loss": 0.4769,
"step": 1156
},
{
"epoch": 5.236006051437216,
"grad_norm": 1.1231337785720825,
"learning_rate": 2.1701001192644067e-07,
"loss": 0.4658,
"step": 1157
},
{
"epoch": 5.24054462934947,
"grad_norm": 0.9594347476959229,
"learning_rate": 2.1439409258197153e-07,
"loss": 0.4819,
"step": 1158
},
{
"epoch": 5.245083207261724,
"grad_norm": 0.7892640233039856,
"learning_rate": 2.117933291187557e-07,
"loss": 0.4498,
"step": 1159
},
{
"epoch": 5.249621785173979,
"grad_norm": 0.8971346616744995,
"learning_rate": 2.092077387824884e-07,
"loss": 0.4792,
"step": 1160
},
{
"epoch": 5.254160363086233,
"grad_norm": 0.7570569515228271,
"learning_rate": 2.0663733871825105e-07,
"loss": 0.4871,
"step": 1161
},
{
"epoch": 5.258698940998487,
"grad_norm": 0.8948609828948975,
"learning_rate": 2.0408214597039928e-07,
"loss": 0.4615,
"step": 1162
},
{
"epoch": 5.263237518910741,
"grad_norm": 0.716673731803894,
"learning_rate": 2.015421774824472e-07,
"loss": 0.4846,
"step": 1163
},
{
"epoch": 5.267776096822995,
"grad_norm": 0.9955589175224304,
"learning_rate": 1.9901745009695772e-07,
"loss": 0.4749,
"step": 1164
},
{
"epoch": 5.27231467473525,
"grad_norm": 0.7813113927841187,
"learning_rate": 1.9650798055543014e-07,
"loss": 0.4854,
"step": 1165
},
{
"epoch": 5.276853252647504,
"grad_norm": 0.7000608444213867,
"learning_rate": 1.9401378549818745e-07,
"loss": 0.4826,
"step": 1166
},
{
"epoch": 5.281391830559758,
"grad_norm": 0.7905756831169128,
"learning_rate": 1.9153488146426897e-07,
"loss": 0.4856,
"step": 1167
},
{
"epoch": 5.285930408472012,
"grad_norm": 0.76811683177948,
"learning_rate": 1.8907128489131732e-07,
"loss": 0.5032,
"step": 1168
},
{
"epoch": 5.290468986384266,
"grad_norm": 0.7297028303146362,
"learning_rate": 1.8662301211547302e-07,
"loss": 0.493,
"step": 1169
},
{
"epoch": 5.295007564296521,
"grad_norm": 0.7701004147529602,
"learning_rate": 1.8419007937126254e-07,
"loss": 0.4593,
"step": 1170
},
{
"epoch": 5.299546142208775,
"grad_norm": 0.7971721887588501,
"learning_rate": 1.8177250279149405e-07,
"loss": 0.4894,
"step": 1171
},
{
"epoch": 5.304084720121029,
"grad_norm": 0.7243720293045044,
"learning_rate": 1.7937029840714716e-07,
"loss": 0.4852,
"step": 1172
},
{
"epoch": 5.308623298033283,
"grad_norm": 0.8683739304542542,
"learning_rate": 1.7698348214726935e-07,
"loss": 0.4563,
"step": 1173
},
{
"epoch": 5.3131618759455375,
"grad_norm": 0.7620447278022766,
"learning_rate": 1.7461206983886952e-07,
"loss": 0.4955,
"step": 1174
},
{
"epoch": 5.317700453857791,
"grad_norm": 0.7649240493774414,
"learning_rate": 1.7225607720681132e-07,
"loss": 0.4707,
"step": 1175
},
{
"epoch": 5.322239031770045,
"grad_norm": 0.9427205920219421,
"learning_rate": 1.6991551987371208e-07,
"loss": 0.4908,
"step": 1176
},
{
"epoch": 5.326777609682299,
"grad_norm": 0.7606921195983887,
"learning_rate": 1.6759041335983606e-07,
"loss": 0.4538,
"step": 1177
},
{
"epoch": 5.331316187594553,
"grad_norm": 0.7385324239730835,
"learning_rate": 1.6528077308299388e-07,
"loss": 0.4719,
"step": 1178
},
{
"epoch": 5.335854765506808,
"grad_norm": 0.8069400787353516,
"learning_rate": 1.6298661435843883e-07,
"loss": 0.4702,
"step": 1179
},
{
"epoch": 5.340393343419062,
"grad_norm": 0.8330691456794739,
"learning_rate": 1.607079523987662e-07,
"loss": 0.4583,
"step": 1180
},
{
"epoch": 5.344931921331316,
"grad_norm": 0.682507336139679,
"learning_rate": 1.5844480231381132e-07,
"loss": 0.4721,
"step": 1181
},
{
"epoch": 5.34947049924357,
"grad_norm": 0.790880560874939,
"learning_rate": 1.5619717911055127e-07,
"loss": 0.4693,
"step": 1182
},
{
"epoch": 5.354009077155824,
"grad_norm": 0.8017518520355225,
"learning_rate": 1.5396509769300294e-07,
"loss": 0.4869,
"step": 1183
},
{
"epoch": 5.358547655068079,
"grad_norm": 0.7758563160896301,
"learning_rate": 1.5174857286212625e-07,
"loss": 0.4819,
"step": 1184
},
{
"epoch": 5.363086232980333,
"grad_norm": 1.1839865446090698,
"learning_rate": 1.4954761931572526e-07,
"loss": 0.5027,
"step": 1185
},
{
"epoch": 5.367624810892587,
"grad_norm": 1.2727587223052979,
"learning_rate": 1.4736225164835e-07,
"loss": 0.4786,
"step": 1186
},
{
"epoch": 5.372163388804841,
"grad_norm": 0.7882137298583984,
"learning_rate": 1.4519248435120093e-07,
"loss": 0.4763,
"step": 1187
},
{
"epoch": 5.376701966717095,
"grad_norm": 0.7846785187721252,
"learning_rate": 1.430383318120318e-07,
"loss": 0.4685,
"step": 1188
},
{
"epoch": 5.38124054462935,
"grad_norm": 0.7354263067245483,
"learning_rate": 1.4089980831505535e-07,
"loss": 0.4718,
"step": 1189
},
{
"epoch": 5.385779122541604,
"grad_norm": 0.7932591438293457,
"learning_rate": 1.3877692804084687e-07,
"loss": 0.4555,
"step": 1190
},
{
"epoch": 5.390317700453858,
"grad_norm": 0.7780957221984863,
"learning_rate": 1.366697050662527e-07,
"loss": 0.5104,
"step": 1191
},
{
"epoch": 5.394856278366112,
"grad_norm": 0.6624619960784912,
"learning_rate": 1.345781533642937e-07,
"loss": 0.5044,
"step": 1192
},
{
"epoch": 5.3993948562783665,
"grad_norm": 0.736815869808197,
"learning_rate": 1.325022868040765e-07,
"loss": 0.4912,
"step": 1193
},
{
"epoch": 5.403933434190621,
"grad_norm": 1.2100348472595215,
"learning_rate": 1.3044211915069715e-07,
"loss": 0.506,
"step": 1194
},
{
"epoch": 5.408472012102875,
"grad_norm": 0.7347081303596497,
"learning_rate": 1.2839766406515296e-07,
"loss": 0.4628,
"step": 1195
},
{
"epoch": 5.413010590015128,
"grad_norm": 0.8266773819923401,
"learning_rate": 1.2636893510425185e-07,
"loss": 0.4952,
"step": 1196
},
{
"epoch": 5.417549167927382,
"grad_norm": 0.7447771430015564,
"learning_rate": 1.2435594572052012e-07,
"loss": 0.4908,
"step": 1197
},
{
"epoch": 5.422087745839637,
"grad_norm": 0.8191699385643005,
"learning_rate": 1.223587092621162e-07,
"loss": 0.4815,
"step": 1198
},
{
"epoch": 5.426626323751891,
"grad_norm": 0.7668866515159607,
"learning_rate": 1.2037723897273895e-07,
"loss": 0.4883,
"step": 1199
},
{
"epoch": 5.431164901664145,
"grad_norm": 1.9579761028289795,
"learning_rate": 1.1841154799154376e-07,
"loss": 0.4937,
"step": 1200
},
{
"epoch": 5.435703479576399,
"grad_norm": 0.7861865758895874,
"learning_rate": 1.16461649353051e-07,
"loss": 0.4899,
"step": 1201
},
{
"epoch": 5.440242057488653,
"grad_norm": 0.7184910774230957,
"learning_rate": 1.1452755598706317e-07,
"loss": 0.4833,
"step": 1202
},
{
"epoch": 5.444780635400908,
"grad_norm": 0.771675169467926,
"learning_rate": 1.1260928071857746e-07,
"loss": 0.4449,
"step": 1203
},
{
"epoch": 5.449319213313162,
"grad_norm": 0.8247400522232056,
"learning_rate": 1.1070683626770163e-07,
"loss": 0.4723,
"step": 1204
},
{
"epoch": 5.453857791225416,
"grad_norm": 0.6917076706886292,
"learning_rate": 1.0882023524956764e-07,
"loss": 0.4666,
"step": 1205
},
{
"epoch": 5.45839636913767,
"grad_norm": 0.7408576607704163,
"learning_rate": 1.0694949017425043e-07,
"loss": 0.4573,
"step": 1206
},
{
"epoch": 5.462934947049924,
"grad_norm": 1.0706077814102173,
"learning_rate": 1.0509461344668486e-07,
"loss": 0.4985,
"step": 1207
},
{
"epoch": 5.467473524962179,
"grad_norm": 0.8229323029518127,
"learning_rate": 1.0325561736658052e-07,
"loss": 0.4774,
"step": 1208
},
{
"epoch": 5.472012102874433,
"grad_norm": 1.0152099132537842,
"learning_rate": 1.0143251412834426e-07,
"loss": 0.5089,
"step": 1209
},
{
"epoch": 5.476550680786687,
"grad_norm": 0.7047984600067139,
"learning_rate": 9.96253158209956e-08,
"loss": 0.4803,
"step": 1210
},
{
"epoch": 5.481089258698941,
"grad_norm": 0.932101309299469,
"learning_rate": 9.783403442808953e-08,
"loss": 0.4897,
"step": 1211
},
{
"epoch": 5.4856278366111955,
"grad_norm": 0.7570725679397583,
"learning_rate": 9.60586818276349e-08,
"loss": 0.4971,
"step": 1212
},
{
"epoch": 5.49016641452345,
"grad_norm": 0.7186459302902222,
"learning_rate": 9.429926979201753e-08,
"loss": 0.4853,
"step": 1213
},
{
"epoch": 5.494704992435704,
"grad_norm": 0.8389477133750916,
"learning_rate": 9.255580998792008e-08,
"loss": 0.4735,
"step": 1214
},
{
"epoch": 5.499243570347957,
"grad_norm": 0.7484689950942993,
"learning_rate": 9.082831397624586e-08,
"loss": 0.4973,
"step": 1215
},
{
"epoch": 5.503782148260212,
"grad_norm": 0.8444059491157532,
"learning_rate": 8.911679321204347e-08,
"loss": 0.4565,
"step": 1216
},
{
"epoch": 5.508320726172466,
"grad_norm": 0.6689127683639526,
"learning_rate": 8.742125904442705e-08,
"loss": 0.4854,
"step": 1217
},
{
"epoch": 5.51285930408472,
"grad_norm": 0.7082827687263489,
"learning_rate": 8.574172271650639e-08,
"loss": 0.4535,
"step": 1218
},
{
"epoch": 5.517397881996974,
"grad_norm": 0.8333298563957214,
"learning_rate": 8.407819536530665e-08,
"loss": 0.459,
"step": 1219
},
{
"epoch": 5.521936459909228,
"grad_norm": 0.7178226709365845,
"learning_rate": 8.243068802169906e-08,
"loss": 0.4836,
"step": 1220
},
{
"epoch": 5.526475037821482,
"grad_norm": 0.8268416523933411,
"learning_rate": 8.079921161032511e-08,
"loss": 0.5056,
"step": 1221
},
{
"epoch": 5.531013615733737,
"grad_norm": 0.6459931135177612,
"learning_rate": 7.918377694952518e-08,
"loss": 0.4748,
"step": 1222
},
{
"epoch": 5.535552193645991,
"grad_norm": 0.7851235270500183,
"learning_rate": 7.758439475126666e-08,
"loss": 0.4837,
"step": 1223
},
{
"epoch": 5.540090771558245,
"grad_norm": 0.8119810223579407,
"learning_rate": 7.600107562107246e-08,
"loss": 0.4597,
"step": 1224
},
{
"epoch": 5.544629349470499,
"grad_norm": 0.7369126081466675,
"learning_rate": 7.443383005795224e-08,
"loss": 0.4659,
"step": 1225
},
{
"epoch": 5.549167927382753,
"grad_norm": 0.7626518607139587,
"learning_rate": 7.288266845433017e-08,
"loss": 0.4847,
"step": 1226
},
{
"epoch": 5.553706505295008,
"grad_norm": 1.6847354173660278,
"learning_rate": 7.134760109597877e-08,
"loss": 0.483,
"step": 1227
},
{
"epoch": 5.558245083207262,
"grad_norm": 0.9522053003311157,
"learning_rate": 6.982863816194785e-08,
"loss": 0.4813,
"step": 1228
},
{
"epoch": 5.562783661119516,
"grad_norm": 0.7725521922111511,
"learning_rate": 6.832578972450099e-08,
"loss": 0.4923,
"step": 1229
},
{
"epoch": 5.56732223903177,
"grad_norm": 0.7546195387840271,
"learning_rate": 6.683906574904364e-08,
"loss": 0.4927,
"step": 1230
},
{
"epoch": 5.5718608169440245,
"grad_norm": 0.928009033203125,
"learning_rate": 6.536847609406178e-08,
"loss": 0.4875,
"step": 1231
},
{
"epoch": 5.576399394856279,
"grad_norm": 0.75600665807724,
"learning_rate": 6.39140305110536e-08,
"loss": 0.4873,
"step": 1232
},
{
"epoch": 5.580937972768533,
"grad_norm": 0.9023406505584717,
"learning_rate": 6.247573864446522e-08,
"loss": 0.4901,
"step": 1233
},
{
"epoch": 5.585476550680786,
"grad_norm": 0.751864492893219,
"learning_rate": 6.105361003162891e-08,
"loss": 0.4785,
"step": 1234
},
{
"epoch": 5.590015128593041,
"grad_norm": 0.7082744836807251,
"learning_rate": 5.964765410269635e-08,
"loss": 0.4884,
"step": 1235
},
{
"epoch": 5.594553706505295,
"grad_norm": 0.6925091743469238,
"learning_rate": 5.825788018057971e-08,
"loss": 0.4762,
"step": 1236
},
{
"epoch": 5.599092284417549,
"grad_norm": 0.7400352358818054,
"learning_rate": 5.688429748088614e-08,
"loss": 0.469,
"step": 1237
},
{
"epoch": 5.603630862329803,
"grad_norm": 0.7220975756645203,
"learning_rate": 5.552691511186065e-08,
"loss": 0.453,
"step": 1238
},
{
"epoch": 5.608169440242057,
"grad_norm": 0.8350204825401306,
"learning_rate": 5.4185742074321645e-08,
"loss": 0.4814,
"step": 1239
},
{
"epoch": 5.612708018154311,
"grad_norm": 0.7439131736755371,
"learning_rate": 5.2860787261605485e-08,
"loss": 0.479,
"step": 1240
},
{
"epoch": 5.617246596066566,
"grad_norm": 0.8671088814735413,
"learning_rate": 5.1552059459503144e-08,
"loss": 0.4823,
"step": 1241
},
{
"epoch": 5.62178517397882,
"grad_norm": 0.7441071271896362,
"learning_rate": 5.0259567346205016e-08,
"loss": 0.4579,
"step": 1242
},
{
"epoch": 5.626323751891074,
"grad_norm": 1.0275096893310547,
"learning_rate": 4.898331949224289e-08,
"loss": 0.5064,
"step": 1243
},
{
"epoch": 5.630862329803328,
"grad_norm": 0.8689181208610535,
"learning_rate": 4.7723324360431644e-08,
"loss": 0.462,
"step": 1244
},
{
"epoch": 5.635400907715582,
"grad_norm": 0.7431532144546509,
"learning_rate": 4.647959030581517e-08,
"loss": 0.4923,
"step": 1245
},
{
"epoch": 5.639939485627837,
"grad_norm": 0.9355483651161194,
"learning_rate": 4.525212557560887e-08,
"loss": 0.459,
"step": 1246
},
{
"epoch": 5.644478063540091,
"grad_norm": 0.7806388735771179,
"learning_rate": 4.404093830914752e-08,
"loss": 0.4567,
"step": 1247
},
{
"epoch": 5.649016641452345,
"grad_norm": 0.7851375937461853,
"learning_rate": 4.284603653782832e-08,
"loss": 0.4641,
"step": 1248
},
{
"epoch": 5.653555219364599,
"grad_norm": 0.826127290725708,
"learning_rate": 4.1667428185060145e-08,
"loss": 0.4824,
"step": 1249
},
{
"epoch": 5.6580937972768535,
"grad_norm": 0.7516127824783325,
"learning_rate": 4.050512106620913e-08,
"loss": 0.4798,
"step": 1250
},
{
"epoch": 5.662632375189108,
"grad_norm": 0.7157534956932068,
"learning_rate": 3.9359122888548984e-08,
"loss": 0.4948,
"step": 1251
},
{
"epoch": 5.667170953101362,
"grad_norm": 0.9715742468833923,
"learning_rate": 3.822944125120742e-08,
"loss": 0.4436,
"step": 1252
},
{
"epoch": 5.671709531013616,
"grad_norm": 0.8162844777107239,
"learning_rate": 3.711608364511787e-08,
"loss": 0.506,
"step": 1253
},
{
"epoch": 5.67624810892587,
"grad_norm": 0.9915447235107422,
"learning_rate": 3.601905745296813e-08,
"loss": 0.5134,
"step": 1254
},
{
"epoch": 5.680786686838124,
"grad_norm": 0.8082160353660583,
"learning_rate": 3.4938369949152616e-08,
"loss": 0.4826,
"step": 1255
},
{
"epoch": 5.685325264750379,
"grad_norm": 2.119804859161377,
"learning_rate": 3.3874028299723805e-08,
"loss": 0.5038,
"step": 1256
},
{
"epoch": 5.689863842662632,
"grad_norm": 0.7752223610877991,
"learning_rate": 3.2826039562343924e-08,
"loss": 0.4629,
"step": 1257
},
{
"epoch": 5.694402420574886,
"grad_norm": 0.7426053285598755,
"learning_rate": 3.1794410686240006e-08,
"loss": 0.4749,
"step": 1258
},
{
"epoch": 5.69894099848714,
"grad_norm": 0.8134835362434387,
"learning_rate": 3.077914851215585e-08,
"loss": 0.4465,
"step": 1259
},
{
"epoch": 5.703479576399395,
"grad_norm": 0.7048810720443726,
"learning_rate": 2.9780259772307362e-08,
"loss": 0.4684,
"step": 1260
},
{
"epoch": 5.708018154311649,
"grad_norm": 0.6710391640663147,
"learning_rate": 2.879775109033811e-08,
"loss": 0.472,
"step": 1261
},
{
"epoch": 5.712556732223903,
"grad_norm": 0.8969804048538208,
"learning_rate": 2.7831628981275504e-08,
"loss": 0.4861,
"step": 1262
},
{
"epoch": 5.717095310136157,
"grad_norm": 1.4039125442504883,
"learning_rate": 2.6881899851487758e-08,
"loss": 0.4634,
"step": 1263
},
{
"epoch": 5.721633888048411,
"grad_norm": 0.6740471124649048,
"learning_rate": 2.5948569998640037e-08,
"loss": 0.49,
"step": 1264
},
{
"epoch": 5.726172465960666,
"grad_norm": 0.671715497970581,
"learning_rate": 2.5031645611654497e-08,
"loss": 0.4782,
"step": 1265
},
{
"epoch": 5.73071104387292,
"grad_norm": 0.7497738003730774,
"learning_rate": 2.413113277066781e-08,
"loss": 0.4396,
"step": 1266
},
{
"epoch": 5.735249621785174,
"grad_norm": 0.739037036895752,
"learning_rate": 2.3247037446992304e-08,
"loss": 0.4561,
"step": 1267
},
{
"epoch": 5.739788199697428,
"grad_norm": 0.7688632607460022,
"learning_rate": 2.2379365503074902e-08,
"loss": 0.4919,
"step": 1268
},
{
"epoch": 5.7443267776096825,
"grad_norm": 1.2213677167892456,
"learning_rate": 2.1528122692459074e-08,
"loss": 0.4927,
"step": 1269
},
{
"epoch": 5.748865355521937,
"grad_norm": 0.7246401309967041,
"learning_rate": 2.0693314659746276e-08,
"loss": 0.4709,
"step": 1270
},
{
"epoch": 5.753403933434191,
"grad_norm": 0.7174831628799438,
"learning_rate": 1.9874946940559305e-08,
"loss": 0.4728,
"step": 1271
},
{
"epoch": 5.757942511346445,
"grad_norm": 0.74196857213974,
"learning_rate": 1.907302496150454e-08,
"loss": 0.4766,
"step": 1272
},
{
"epoch": 5.762481089258699,
"grad_norm": 0.8071885108947754,
"learning_rate": 1.8287554040136714e-08,
"loss": 0.4954,
"step": 1273
},
{
"epoch": 5.7670196671709535,
"grad_norm": 0.9311023950576782,
"learning_rate": 1.751853938492337e-08,
"loss": 0.4315,
"step": 1274
},
{
"epoch": 5.771558245083208,
"grad_norm": 0.8586441278457642,
"learning_rate": 1.6765986095209906e-08,
"loss": 0.5136,
"step": 1275
},
{
"epoch": 5.776096822995461,
"grad_norm": 0.7332836985588074,
"learning_rate": 1.602989916118708e-08,
"loss": 0.5048,
"step": 1276
},
{
"epoch": 5.780635400907715,
"grad_norm": 1.2354652881622314,
"learning_rate": 1.531028346385577e-08,
"loss": 0.4652,
"step": 1277
},
{
"epoch": 5.785173978819969,
"grad_norm": 0.6708296537399292,
"learning_rate": 1.4607143774997557e-08,
"loss": 0.4953,
"step": 1278
},
{
"epoch": 5.789712556732224,
"grad_norm": 0.8077536821365356,
"learning_rate": 1.3920484757139751e-08,
"loss": 0.5044,
"step": 1279
},
{
"epoch": 5.794251134644478,
"grad_norm": 0.9781695604324341,
"learning_rate": 1.3250310963527358e-08,
"loss": 0.5122,
"step": 1280
},
{
"epoch": 5.798789712556732,
"grad_norm": 0.7283720374107361,
"learning_rate": 1.2596626838090875e-08,
"loss": 0.4435,
"step": 1281
},
{
"epoch": 5.803328290468986,
"grad_norm": 0.7447962760925293,
"learning_rate": 1.1959436715417993e-08,
"loss": 0.4749,
"step": 1282
},
{
"epoch": 5.80786686838124,
"grad_norm": 0.7438759207725525,
"learning_rate": 1.1338744820724168e-08,
"loss": 0.4774,
"step": 1283
},
{
"epoch": 5.812405446293495,
"grad_norm": 0.7419090270996094,
"learning_rate": 1.0734555269825141e-08,
"loss": 0.4876,
"step": 1284
},
{
"epoch": 5.816944024205749,
"grad_norm": 0.8018777966499329,
"learning_rate": 1.0146872069109748e-08,
"loss": 0.48,
"step": 1285
},
{
"epoch": 5.821482602118003,
"grad_norm": 0.7206845879554749,
"learning_rate": 9.575699115512149e-09,
"loss": 0.4402,
"step": 1286
},
{
"epoch": 5.826021180030257,
"grad_norm": 0.8025342226028442,
"learning_rate": 9.021040196487419e-09,
"loss": 0.5233,
"step": 1287
},
{
"epoch": 5.8305597579425115,
"grad_norm": 0.7229884266853333,
"learning_rate": 8.482898989985721e-09,
"loss": 0.5087,
"step": 1288
},
{
"epoch": 5.835098335854766,
"grad_norm": 0.809455931186676,
"learning_rate": 7.961279064428452e-09,
"loss": 0.4836,
"step": 1289
},
{
"epoch": 5.83963691376702,
"grad_norm": 0.8334141969680786,
"learning_rate": 7.456183878683243e-09,
"loss": 0.4724,
"step": 1290
},
{
"epoch": 5.844175491679274,
"grad_norm": 2.202322244644165,
"learning_rate": 6.967616782043163e-09,
"loss": 0.48,
"step": 1291
},
{
"epoch": 5.848714069591528,
"grad_norm": 0.7208507061004639,
"learning_rate": 6.495581014202557e-09,
"loss": 0.4551,
"step": 1292
},
{
"epoch": 5.8532526475037825,
"grad_norm": 0.8301936984062195,
"learning_rate": 6.040079705235957e-09,
"loss": 0.4666,
"step": 1293
},
{
"epoch": 5.857791225416037,
"grad_norm": 0.7231348752975464,
"learning_rate": 5.601115875578933e-09,
"loss": 0.4478,
"step": 1294
},
{
"epoch": 5.86232980332829,
"grad_norm": 0.740260124206543,
"learning_rate": 5.178692436005883e-09,
"loss": 0.4785,
"step": 1295
},
{
"epoch": 5.866868381240545,
"grad_norm": 0.7141087651252747,
"learning_rate": 4.772812187611719e-09,
"loss": 0.5093,
"step": 1296
},
{
"epoch": 5.871406959152798,
"grad_norm": 0.6469274759292603,
"learning_rate": 4.383477821793547e-09,
"loss": 0.511,
"step": 1297
},
{
"epoch": 5.875945537065053,
"grad_norm": 0.7152246832847595,
"learning_rate": 4.010691920232623e-09,
"loss": 0.5036,
"step": 1298
},
{
"epoch": 5.880484114977307,
"grad_norm": 0.8838086128234863,
"learning_rate": 3.6544569548763175e-09,
"loss": 0.4956,
"step": 1299
},
{
"epoch": 5.885022692889561,
"grad_norm": 0.7219541072845459,
"learning_rate": 3.3147752879236773e-09,
"loss": 0.5097,
"step": 1300
},
{
"epoch": 5.889561270801815,
"grad_norm": 1.4032868146896362,
"learning_rate": 2.991649171807942e-09,
"loss": 0.4439,
"step": 1301
},
{
"epoch": 5.8940998487140694,
"grad_norm": 1.0518102645874023,
"learning_rate": 2.6850807491823893e-09,
"loss": 0.4909,
"step": 1302
},
{
"epoch": 5.898638426626324,
"grad_norm": 0.7985177636146545,
"learning_rate": 2.3950720529059003e-09,
"loss": 0.4639,
"step": 1303
},
{
"epoch": 5.903177004538578,
"grad_norm": 0.8552533984184265,
"learning_rate": 2.1216250060296374e-09,
"loss": 0.4552,
"step": 1304
},
{
"epoch": 5.907715582450832,
"grad_norm": 0.711228609085083,
"learning_rate": 1.8647414217848325e-09,
"loss": 0.4411,
"step": 1305
},
{
"epoch": 5.912254160363086,
"grad_norm": 0.7168917655944824,
"learning_rate": 1.6244230035694642e-09,
"loss": 0.4903,
"step": 1306
},
{
"epoch": 5.9167927382753405,
"grad_norm": 0.7238544225692749,
"learning_rate": 1.400671344937987e-09,
"loss": 0.4901,
"step": 1307
},
{
"epoch": 5.921331316187595,
"grad_norm": 1.0520352125167847,
"learning_rate": 1.1934879295905089e-09,
"loss": 0.4943,
"step": 1308
},
{
"epoch": 5.925869894099849,
"grad_norm": 0.7716150283813477,
"learning_rate": 1.0028741313627965e-09,
"loss": 0.4732,
"step": 1309
},
{
"epoch": 5.930408472012103,
"grad_norm": 0.8260288238525391,
"learning_rate": 8.28831214217396e-10,
"loss": 0.4676,
"step": 1310
},
{
"epoch": 5.934947049924357,
"grad_norm": 0.69765704870224,
"learning_rate": 6.713603322347495e-10,
"loss": 0.489,
"step": 1311
},
{
"epoch": 5.9394856278366115,
"grad_norm": 1.1246147155761719,
"learning_rate": 5.304625296065346e-10,
"loss": 0.4862,
"step": 1312
},
{
"epoch": 5.944024205748866,
"grad_norm": 0.7668100595474243,
"learning_rate": 4.0613874062761515e-10,
"loss": 0.4825,
"step": 1313
},
{
"epoch": 5.94856278366112,
"grad_norm": 0.8186943531036377,
"learning_rate": 2.983897896910448e-10,
"loss": 0.4983,
"step": 1314
},
{
"epoch": 5.953101361573374,
"grad_norm": 0.7153812050819397,
"learning_rate": 2.0721639128085113e-10,
"loss": 0.4683,
"step": 1315
},
{
"epoch": 5.957639939485627,
"grad_norm": 0.9169692397117615,
"learning_rate": 1.3261914996953728e-10,
"loss": 0.4744,
"step": 1316
},
{
"epoch": 5.962178517397882,
"grad_norm": 0.7419823408126831,
"learning_rate": 7.459856041169833e-11,
"loss": 0.5012,
"step": 1317
},
{
"epoch": 5.966717095310136,
"grad_norm": 0.8549036383628845,
"learning_rate": 3.315500734318855e-11,
"loss": 0.4881,
"step": 1318
},
{
"epoch": 5.97125567322239,
"grad_norm": 0.6958428621292114,
"learning_rate": 8.288765576403013e-12,
"loss": 0.4775,
"step": 1319
},
{
"epoch": 5.975794251134644,
"grad_norm": 0.7530087828636169,
"learning_rate": 0.0,
"loss": 0.4666,
"step": 1320
}
],
"logging_steps": 1,
"max_steps": 1320,
"num_input_tokens_seen": 0,
"num_train_epochs": 6,
"save_steps": 220,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.99618736616217e+19,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}