Delta-Vector's picture
Training in progress, step 1260, checkpoint
a9df801 verified
raw
history blame
221 kB
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9940546967895365,
"eval_steps": 105,
"global_step": 1260,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0023781212841854932,
"grad_norm": 32.74397118558861,
"learning_rate": 5.000000000000001e-07,
"loss": 2.9478,
"step": 1
},
{
"epoch": 0.0023781212841854932,
"eval_loss": 3.373392343521118,
"eval_runtime": 78.9756,
"eval_samples_per_second": 8.091,
"eval_steps_per_second": 1.013,
"step": 1
},
{
"epoch": 0.0047562425683709865,
"grad_norm": 37.080911592721954,
"learning_rate": 1.0000000000000002e-06,
"loss": 2.8264,
"step": 2
},
{
"epoch": 0.007134363852556481,
"grad_norm": 31.107267997702266,
"learning_rate": 1.5e-06,
"loss": 2.9936,
"step": 3
},
{
"epoch": 0.009512485136741973,
"grad_norm": 27.574905774161167,
"learning_rate": 2.0000000000000003e-06,
"loss": 2.8944,
"step": 4
},
{
"epoch": 0.011890606420927468,
"grad_norm": 13.643957484299273,
"learning_rate": 2.5e-06,
"loss": 2.7687,
"step": 5
},
{
"epoch": 0.014268727705112961,
"grad_norm": 15.077028980653411,
"learning_rate": 3e-06,
"loss": 2.6623,
"step": 6
},
{
"epoch": 0.016646848989298454,
"grad_norm": 14.569557474559408,
"learning_rate": 3.5e-06,
"loss": 2.9007,
"step": 7
},
{
"epoch": 0.019024970273483946,
"grad_norm": 13.894984550517007,
"learning_rate": 4.000000000000001e-06,
"loss": 2.8359,
"step": 8
},
{
"epoch": 0.02140309155766944,
"grad_norm": 13.546442233583257,
"learning_rate": 4.5e-06,
"loss": 2.7583,
"step": 9
},
{
"epoch": 0.023781212841854936,
"grad_norm": 11.166476039936938,
"learning_rate": 5e-06,
"loss": 2.518,
"step": 10
},
{
"epoch": 0.026159334126040427,
"grad_norm": 7.787448179397784,
"learning_rate": 5.500000000000001e-06,
"loss": 2.6494,
"step": 11
},
{
"epoch": 0.028537455410225922,
"grad_norm": 4.72349822440695,
"learning_rate": 6e-06,
"loss": 2.4022,
"step": 12
},
{
"epoch": 0.030915576694411414,
"grad_norm": 4.100722460414476,
"learning_rate": 6.5000000000000004e-06,
"loss": 2.3933,
"step": 13
},
{
"epoch": 0.03329369797859691,
"grad_norm": 4.193151112965372,
"learning_rate": 7e-06,
"loss": 2.5468,
"step": 14
},
{
"epoch": 0.0356718192627824,
"grad_norm": 5.502246954578136,
"learning_rate": 7.500000000000001e-06,
"loss": 2.5126,
"step": 15
},
{
"epoch": 0.03804994054696789,
"grad_norm": 5.716937946349337,
"learning_rate": 8.000000000000001e-06,
"loss": 2.6761,
"step": 16
},
{
"epoch": 0.04042806183115339,
"grad_norm": 3.008867017303434,
"learning_rate": 8.5e-06,
"loss": 2.4264,
"step": 17
},
{
"epoch": 0.04280618311533888,
"grad_norm": 3.6016120293217178,
"learning_rate": 9e-06,
"loss": 2.3836,
"step": 18
},
{
"epoch": 0.04518430439952437,
"grad_norm": 2.1431386543975908,
"learning_rate": 9.5e-06,
"loss": 2.4879,
"step": 19
},
{
"epoch": 0.04756242568370987,
"grad_norm": 3.838539096237921,
"learning_rate": 1e-05,
"loss": 2.4116,
"step": 20
},
{
"epoch": 0.04994054696789536,
"grad_norm": 1.9119200890173822,
"learning_rate": 1.0500000000000001e-05,
"loss": 2.6716,
"step": 21
},
{
"epoch": 0.052318668252080855,
"grad_norm": 2.0853737807318904,
"learning_rate": 1.1000000000000001e-05,
"loss": 2.4523,
"step": 22
},
{
"epoch": 0.054696789536266346,
"grad_norm": 1.6109539551135528,
"learning_rate": 1.15e-05,
"loss": 2.3086,
"step": 23
},
{
"epoch": 0.057074910820451845,
"grad_norm": 2.575933824126331,
"learning_rate": 1.2e-05,
"loss": 2.5576,
"step": 24
},
{
"epoch": 0.059453032104637336,
"grad_norm": 1.4827660467902501,
"learning_rate": 1.25e-05,
"loss": 2.264,
"step": 25
},
{
"epoch": 0.06183115338882283,
"grad_norm": 2.1022593224903128,
"learning_rate": 1.3000000000000001e-05,
"loss": 2.8781,
"step": 26
},
{
"epoch": 0.06420927467300833,
"grad_norm": 2.756765764204405,
"learning_rate": 1.3500000000000001e-05,
"loss": 2.5071,
"step": 27
},
{
"epoch": 0.06658739595719382,
"grad_norm": 1.6249455774563617,
"learning_rate": 1.4e-05,
"loss": 2.5085,
"step": 28
},
{
"epoch": 0.06896551724137931,
"grad_norm": 1.8653671914008603,
"learning_rate": 1.45e-05,
"loss": 2.4368,
"step": 29
},
{
"epoch": 0.0713436385255648,
"grad_norm": 1.867632087572305,
"learning_rate": 1.5000000000000002e-05,
"loss": 2.1204,
"step": 30
},
{
"epoch": 0.07372175980975029,
"grad_norm": 1.2367329590346625,
"learning_rate": 1.55e-05,
"loss": 2.1118,
"step": 31
},
{
"epoch": 0.07609988109393578,
"grad_norm": 1.5969207000997636,
"learning_rate": 1.6000000000000003e-05,
"loss": 2.4747,
"step": 32
},
{
"epoch": 0.07847800237812129,
"grad_norm": 1.5075133034220278,
"learning_rate": 1.65e-05,
"loss": 2.3349,
"step": 33
},
{
"epoch": 0.08085612366230678,
"grad_norm": 1.4803692479229955,
"learning_rate": 1.7e-05,
"loss": 2.3915,
"step": 34
},
{
"epoch": 0.08323424494649227,
"grad_norm": 2.692949771759104,
"learning_rate": 1.7500000000000002e-05,
"loss": 2.6397,
"step": 35
},
{
"epoch": 0.08561236623067776,
"grad_norm": 1.3636512770329847,
"learning_rate": 1.8e-05,
"loss": 2.2944,
"step": 36
},
{
"epoch": 0.08799048751486326,
"grad_norm": 1.8310086696195464,
"learning_rate": 1.8500000000000002e-05,
"loss": 2.2614,
"step": 37
},
{
"epoch": 0.09036860879904875,
"grad_norm": 1.7765027708264853,
"learning_rate": 1.9e-05,
"loss": 2.3579,
"step": 38
},
{
"epoch": 0.09274673008323424,
"grad_norm": 1.4484769960901491,
"learning_rate": 1.95e-05,
"loss": 2.4548,
"step": 39
},
{
"epoch": 0.09512485136741974,
"grad_norm": 3.1520205275209414,
"learning_rate": 2e-05,
"loss": 2.5208,
"step": 40
},
{
"epoch": 0.09750297265160524,
"grad_norm": 1.5897739849482102,
"learning_rate": 1.9999981652287733e-05,
"loss": 2.4216,
"step": 41
},
{
"epoch": 0.09988109393579073,
"grad_norm": 2.4520591326987495,
"learning_rate": 1.999992660921826e-05,
"loss": 2.2326,
"step": 42
},
{
"epoch": 0.10225921521997622,
"grad_norm": 1.4817926319377914,
"learning_rate": 1.999983487099356e-05,
"loss": 2.541,
"step": 43
},
{
"epoch": 0.10463733650416171,
"grad_norm": 2.4022755616863956,
"learning_rate": 1.999970643795027e-05,
"loss": 2.3645,
"step": 44
},
{
"epoch": 0.1070154577883472,
"grad_norm": 2.476073007712477,
"learning_rate": 1.9999541310559686e-05,
"loss": 2.5051,
"step": 45
},
{
"epoch": 0.10939357907253269,
"grad_norm": 3.428213096316913,
"learning_rate": 1.9999339489427746e-05,
"loss": 2.3605,
"step": 46
},
{
"epoch": 0.1117717003567182,
"grad_norm": 2.464783346708793,
"learning_rate": 1.9999100975295046e-05,
"loss": 2.3785,
"step": 47
},
{
"epoch": 0.11414982164090369,
"grad_norm": 2.0686444585541754,
"learning_rate": 1.999882576903682e-05,
"loss": 2.6035,
"step": 48
},
{
"epoch": 0.11652794292508918,
"grad_norm": 2.30832669621963,
"learning_rate": 1.9998513871662945e-05,
"loss": 2.3982,
"step": 49
},
{
"epoch": 0.11890606420927467,
"grad_norm": 1.7443884357427357,
"learning_rate": 1.9998165284317944e-05,
"loss": 2.2344,
"step": 50
},
{
"epoch": 0.12128418549346016,
"grad_norm": 1.862498116501275,
"learning_rate": 1.999778000828098e-05,
"loss": 2.0631,
"step": 51
},
{
"epoch": 0.12366230677764566,
"grad_norm": 1.3054707043181313,
"learning_rate": 1.9997358044965833e-05,
"loss": 2.4434,
"step": 52
},
{
"epoch": 0.12604042806183116,
"grad_norm": 2.6645481874919583,
"learning_rate": 1.9996899395920915e-05,
"loss": 2.4647,
"step": 53
},
{
"epoch": 0.12841854934601665,
"grad_norm": 1.6884769054479885,
"learning_rate": 1.999640406282926e-05,
"loss": 2.2864,
"step": 54
},
{
"epoch": 0.13079667063020214,
"grad_norm": 2.281038989939936,
"learning_rate": 1.9995872047508516e-05,
"loss": 2.2999,
"step": 55
},
{
"epoch": 0.13317479191438764,
"grad_norm": 1.8457967068582515,
"learning_rate": 1.9995303351910934e-05,
"loss": 2.5117,
"step": 56
},
{
"epoch": 0.13555291319857313,
"grad_norm": 1.593901860141829,
"learning_rate": 1.9994697978123363e-05,
"loss": 2.1696,
"step": 57
},
{
"epoch": 0.13793103448275862,
"grad_norm": 2.168613470386471,
"learning_rate": 1.9994055928367256e-05,
"loss": 2.6852,
"step": 58
},
{
"epoch": 0.1403091557669441,
"grad_norm": 1.5398971278043392,
"learning_rate": 1.999337720499863e-05,
"loss": 2.396,
"step": 59
},
{
"epoch": 0.1426872770511296,
"grad_norm": 1.7618365641290346,
"learning_rate": 1.99926618105081e-05,
"loss": 2.047,
"step": 60
},
{
"epoch": 0.1450653983353151,
"grad_norm": 1.3935889380673343,
"learning_rate": 1.9991909747520835e-05,
"loss": 2.1071,
"step": 61
},
{
"epoch": 0.14744351961950058,
"grad_norm": 1.3063537354837544,
"learning_rate": 1.999112101879656e-05,
"loss": 2.3992,
"step": 62
},
{
"epoch": 0.14982164090368608,
"grad_norm": 1.025531260937785,
"learning_rate": 1.9990295627229544e-05,
"loss": 2.4764,
"step": 63
},
{
"epoch": 0.15219976218787157,
"grad_norm": 1.0662434476421614,
"learning_rate": 1.99894335758486e-05,
"loss": 2.0608,
"step": 64
},
{
"epoch": 0.1545778834720571,
"grad_norm": 1.335312659171346,
"learning_rate": 1.9988534867817065e-05,
"loss": 2.0345,
"step": 65
},
{
"epoch": 0.15695600475624258,
"grad_norm": 1.7866360477276542,
"learning_rate": 1.9987599506432785e-05,
"loss": 2.4781,
"step": 66
},
{
"epoch": 0.15933412604042807,
"grad_norm": 1.3661388854405736,
"learning_rate": 1.9986627495128105e-05,
"loss": 2.3895,
"step": 67
},
{
"epoch": 0.16171224732461356,
"grad_norm": 1.69413799763372,
"learning_rate": 1.9985618837469864e-05,
"loss": 2.415,
"step": 68
},
{
"epoch": 0.16409036860879905,
"grad_norm": 1.6458921632697698,
"learning_rate": 1.998457353715938e-05,
"loss": 2.4017,
"step": 69
},
{
"epoch": 0.16646848989298454,
"grad_norm": 1.705150979387567,
"learning_rate": 1.998349159803241e-05,
"loss": 2.317,
"step": 70
},
{
"epoch": 0.16884661117717004,
"grad_norm": 1.4786622132550975,
"learning_rate": 1.9982373024059195e-05,
"loss": 2.4046,
"step": 71
},
{
"epoch": 0.17122473246135553,
"grad_norm": 1.1257378401253821,
"learning_rate": 1.998121781934438e-05,
"loss": 2.196,
"step": 72
},
{
"epoch": 0.17360285374554102,
"grad_norm": 1.2450243917414825,
"learning_rate": 1.9980025988127037e-05,
"loss": 2.1235,
"step": 73
},
{
"epoch": 0.1759809750297265,
"grad_norm": 1.2574977788035384,
"learning_rate": 1.9978797534780646e-05,
"loss": 2.4466,
"step": 74
},
{
"epoch": 0.178359096313912,
"grad_norm": 1.390309850165232,
"learning_rate": 1.9977532463813064e-05,
"loss": 2.4469,
"step": 75
},
{
"epoch": 0.1807372175980975,
"grad_norm": 1.2146735833909619,
"learning_rate": 1.9976230779866527e-05,
"loss": 2.3705,
"step": 76
},
{
"epoch": 0.18311533888228299,
"grad_norm": 1.26356031715395,
"learning_rate": 1.9974892487717613e-05,
"loss": 2.4926,
"step": 77
},
{
"epoch": 0.18549346016646848,
"grad_norm": 1.3934309027656608,
"learning_rate": 1.997351759227725e-05,
"loss": 2.47,
"step": 78
},
{
"epoch": 0.187871581450654,
"grad_norm": 1.129998342751621,
"learning_rate": 1.9972106098590665e-05,
"loss": 2.3718,
"step": 79
},
{
"epoch": 0.1902497027348395,
"grad_norm": 1.3957720808228478,
"learning_rate": 1.9970658011837404e-05,
"loss": 2.2057,
"step": 80
},
{
"epoch": 0.19262782401902498,
"grad_norm": 1.5835508346410572,
"learning_rate": 1.9969173337331283e-05,
"loss": 2.2551,
"step": 81
},
{
"epoch": 0.19500594530321047,
"grad_norm": 1.0726363834452401,
"learning_rate": 1.996765208052037e-05,
"loss": 1.9962,
"step": 82
},
{
"epoch": 0.19738406658739596,
"grad_norm": 1.5380248440103288,
"learning_rate": 1.9966094246986983e-05,
"loss": 2.1986,
"step": 83
},
{
"epoch": 0.19976218787158145,
"grad_norm": 1.2953387209833067,
"learning_rate": 1.9964499842447665e-05,
"loss": 2.5842,
"step": 84
},
{
"epoch": 0.20214030915576695,
"grad_norm": 1.0712033116668103,
"learning_rate": 1.9962868872753144e-05,
"loss": 2.1298,
"step": 85
},
{
"epoch": 0.20451843043995244,
"grad_norm": 1.095761902776689,
"learning_rate": 1.996120134388834e-05,
"loss": 2.3641,
"step": 86
},
{
"epoch": 0.20689655172413793,
"grad_norm": 1.186679631328553,
"learning_rate": 1.995949726197231e-05,
"loss": 2.4801,
"step": 87
},
{
"epoch": 0.20927467300832342,
"grad_norm": 1.1887498108170933,
"learning_rate": 1.9957756633258264e-05,
"loss": 2.2866,
"step": 88
},
{
"epoch": 0.2116527942925089,
"grad_norm": 1.0909023774872124,
"learning_rate": 1.9955979464133515e-05,
"loss": 2.2916,
"step": 89
},
{
"epoch": 0.2140309155766944,
"grad_norm": 1.1077175360558418,
"learning_rate": 1.995416576111945e-05,
"loss": 2.1077,
"step": 90
},
{
"epoch": 0.2164090368608799,
"grad_norm": 1.1526064687436712,
"learning_rate": 1.9952315530871537e-05,
"loss": 2.2723,
"step": 91
},
{
"epoch": 0.21878715814506539,
"grad_norm": 1.1818210038912647,
"learning_rate": 1.9950428780179274e-05,
"loss": 2.2338,
"step": 92
},
{
"epoch": 0.2211652794292509,
"grad_norm": 1.164942154271255,
"learning_rate": 1.994850551596617e-05,
"loss": 2.3817,
"step": 93
},
{
"epoch": 0.2235434007134364,
"grad_norm": 1.339398993177121,
"learning_rate": 1.9946545745289727e-05,
"loss": 2.5508,
"step": 94
},
{
"epoch": 0.2259215219976219,
"grad_norm": 1.3267763052855093,
"learning_rate": 1.9944549475341404e-05,
"loss": 2.247,
"step": 95
},
{
"epoch": 0.22829964328180738,
"grad_norm": 1.1953250811556597,
"learning_rate": 1.99425167134466e-05,
"loss": 2.3373,
"step": 96
},
{
"epoch": 0.23067776456599287,
"grad_norm": 1.4321452409301854,
"learning_rate": 1.9940447467064624e-05,
"loss": 2.4776,
"step": 97
},
{
"epoch": 0.23305588585017836,
"grad_norm": 1.0224444212683161,
"learning_rate": 1.9938341743788658e-05,
"loss": 2.1837,
"step": 98
},
{
"epoch": 0.23543400713436385,
"grad_norm": 1.0977574950238398,
"learning_rate": 1.9936199551345744e-05,
"loss": 2.1478,
"step": 99
},
{
"epoch": 0.23781212841854935,
"grad_norm": 1.0660069054078747,
"learning_rate": 1.9934020897596752e-05,
"loss": 2.2816,
"step": 100
},
{
"epoch": 0.24019024970273484,
"grad_norm": 1.0312993672336248,
"learning_rate": 1.9931805790536342e-05,
"loss": 2.2468,
"step": 101
},
{
"epoch": 0.24256837098692033,
"grad_norm": 1.1278898252066067,
"learning_rate": 1.9929554238292944e-05,
"loss": 2.0526,
"step": 102
},
{
"epoch": 0.24494649227110582,
"grad_norm": 1.288343002944789,
"learning_rate": 1.992726624912872e-05,
"loss": 2.512,
"step": 103
},
{
"epoch": 0.2473246135552913,
"grad_norm": 1.1840020386119305,
"learning_rate": 1.992494183143955e-05,
"loss": 2.6356,
"step": 104
},
{
"epoch": 0.2497027348394768,
"grad_norm": 1.1014678408276726,
"learning_rate": 1.9922580993754985e-05,
"loss": 2.3521,
"step": 105
},
{
"epoch": 0.2497027348394768,
"eval_loss": 2.5365779399871826,
"eval_runtime": 66.0796,
"eval_samples_per_second": 9.67,
"eval_steps_per_second": 1.211,
"step": 105
},
{
"epoch": 0.2520808561236623,
"grad_norm": 0.9957844757920508,
"learning_rate": 1.9920183744738208e-05,
"loss": 2.355,
"step": 106
},
{
"epoch": 0.2544589774078478,
"grad_norm": 1.0070598447313825,
"learning_rate": 1.9917750093186036e-05,
"loss": 2.1747,
"step": 107
},
{
"epoch": 0.2568370986920333,
"grad_norm": 1.2463453868295562,
"learning_rate": 1.9915280048028853e-05,
"loss": 2.4131,
"step": 108
},
{
"epoch": 0.25921521997621877,
"grad_norm": 1.158558292534161,
"learning_rate": 1.9912773618330595e-05,
"loss": 2.4527,
"step": 109
},
{
"epoch": 0.2615933412604043,
"grad_norm": 1.1875643459332377,
"learning_rate": 1.9910230813288713e-05,
"loss": 2.1523,
"step": 110
},
{
"epoch": 0.26397146254458975,
"grad_norm": 0.892269173897758,
"learning_rate": 1.9907651642234138e-05,
"loss": 1.9606,
"step": 111
},
{
"epoch": 0.26634958382877527,
"grad_norm": 1.181952902180908,
"learning_rate": 1.9905036114631247e-05,
"loss": 2.3201,
"step": 112
},
{
"epoch": 0.26872770511296074,
"grad_norm": 0.9689153704257877,
"learning_rate": 1.990238424007783e-05,
"loss": 2.2329,
"step": 113
},
{
"epoch": 0.27110582639714625,
"grad_norm": 1.3665918769424286,
"learning_rate": 1.989969602830505e-05,
"loss": 2.2387,
"step": 114
},
{
"epoch": 0.2734839476813318,
"grad_norm": 1.0478434719151144,
"learning_rate": 1.9896971489177417e-05,
"loss": 2.2798,
"step": 115
},
{
"epoch": 0.27586206896551724,
"grad_norm": 1.5752154316391798,
"learning_rate": 1.9894210632692745e-05,
"loss": 2.201,
"step": 116
},
{
"epoch": 0.27824019024970276,
"grad_norm": 1.0264277011384757,
"learning_rate": 1.9891413468982112e-05,
"loss": 2.2756,
"step": 117
},
{
"epoch": 0.2806183115338882,
"grad_norm": 1.1750703393359614,
"learning_rate": 1.988858000830983e-05,
"loss": 2.1907,
"step": 118
},
{
"epoch": 0.28299643281807374,
"grad_norm": 0.9456957190962577,
"learning_rate": 1.9885710261073402e-05,
"loss": 2.2993,
"step": 119
},
{
"epoch": 0.2853745541022592,
"grad_norm": 1.37591692336223,
"learning_rate": 1.9882804237803487e-05,
"loss": 2.0751,
"step": 120
},
{
"epoch": 0.2877526753864447,
"grad_norm": 0.991181263305241,
"learning_rate": 1.9879861949163863e-05,
"loss": 2.1946,
"step": 121
},
{
"epoch": 0.2901307966706302,
"grad_norm": 1.2826616603092615,
"learning_rate": 1.9876883405951378e-05,
"loss": 2.3084,
"step": 122
},
{
"epoch": 0.2925089179548157,
"grad_norm": 1.3162982027829009,
"learning_rate": 1.987386861909593e-05,
"loss": 2.294,
"step": 123
},
{
"epoch": 0.29488703923900117,
"grad_norm": 1.086311999313279,
"learning_rate": 1.98708175996604e-05,
"loss": 2.3025,
"step": 124
},
{
"epoch": 0.2972651605231867,
"grad_norm": 1.10683170372015,
"learning_rate": 1.986773035884064e-05,
"loss": 2.3447,
"step": 125
},
{
"epoch": 0.29964328180737215,
"grad_norm": 1.090568761480393,
"learning_rate": 1.9864606907965407e-05,
"loss": 2.4104,
"step": 126
},
{
"epoch": 0.30202140309155767,
"grad_norm": 1.4024759238343605,
"learning_rate": 1.986144725849634e-05,
"loss": 2.298,
"step": 127
},
{
"epoch": 0.30439952437574314,
"grad_norm": 0.9324914520062791,
"learning_rate": 1.9858251422027903e-05,
"loss": 2.1123,
"step": 128
},
{
"epoch": 0.30677764565992865,
"grad_norm": 1.3818136151492852,
"learning_rate": 1.9855019410287355e-05,
"loss": 2.2786,
"step": 129
},
{
"epoch": 0.3091557669441142,
"grad_norm": 0.9879756737720099,
"learning_rate": 1.98517512351347e-05,
"loss": 2.2735,
"step": 130
},
{
"epoch": 0.31153388822829964,
"grad_norm": 1.4107106057474024,
"learning_rate": 1.9848446908562647e-05,
"loss": 2.2421,
"step": 131
},
{
"epoch": 0.31391200951248516,
"grad_norm": 0.978862094447652,
"learning_rate": 1.9845106442696563e-05,
"loss": 2.4152,
"step": 132
},
{
"epoch": 0.3162901307966706,
"grad_norm": 1.3714074038447606,
"learning_rate": 1.9841729849794427e-05,
"loss": 2.4567,
"step": 133
},
{
"epoch": 0.31866825208085614,
"grad_norm": 1.030641093673837,
"learning_rate": 1.983831714224679e-05,
"loss": 2.3015,
"step": 134
},
{
"epoch": 0.3210463733650416,
"grad_norm": 1.1744699755999302,
"learning_rate": 1.9834868332576727e-05,
"loss": 2.2878,
"step": 135
},
{
"epoch": 0.3234244946492271,
"grad_norm": 0.9733999816490441,
"learning_rate": 1.9831383433439798e-05,
"loss": 2.1571,
"step": 136
},
{
"epoch": 0.3258026159334126,
"grad_norm": 1.0470367999253474,
"learning_rate": 1.982786245762398e-05,
"loss": 2.0943,
"step": 137
},
{
"epoch": 0.3281807372175981,
"grad_norm": 1.0748276455064096,
"learning_rate": 1.9824305418049645e-05,
"loss": 2.4156,
"step": 138
},
{
"epoch": 0.33055885850178357,
"grad_norm": 1.0220509349947084,
"learning_rate": 1.9820712327769503e-05,
"loss": 2.1898,
"step": 139
},
{
"epoch": 0.3329369797859691,
"grad_norm": 0.9811166423920332,
"learning_rate": 1.9817083199968552e-05,
"loss": 2.3449,
"step": 140
},
{
"epoch": 0.33531510107015455,
"grad_norm": 1.0664757695722766,
"learning_rate": 1.9813418047964025e-05,
"loss": 2.1514,
"step": 141
},
{
"epoch": 0.3376932223543401,
"grad_norm": 1.1228830278366924,
"learning_rate": 1.9809716885205363e-05,
"loss": 2.3371,
"step": 142
},
{
"epoch": 0.3400713436385256,
"grad_norm": 1.0703957613617774,
"learning_rate": 1.980597972527413e-05,
"loss": 2.2577,
"step": 143
},
{
"epoch": 0.34244946492271106,
"grad_norm": 0.9971842999532138,
"learning_rate": 1.9802206581883992e-05,
"loss": 2.2048,
"step": 144
},
{
"epoch": 0.3448275862068966,
"grad_norm": 0.9969712850303254,
"learning_rate": 1.979839746888067e-05,
"loss": 2.1725,
"step": 145
},
{
"epoch": 0.34720570749108204,
"grad_norm": 0.9782490093980141,
"learning_rate": 1.979455240024186e-05,
"loss": 2.1598,
"step": 146
},
{
"epoch": 0.34958382877526756,
"grad_norm": 1.1595035293528873,
"learning_rate": 1.97906713900772e-05,
"loss": 2.1812,
"step": 147
},
{
"epoch": 0.351961950059453,
"grad_norm": 1.0488323565717943,
"learning_rate": 1.9786754452628226e-05,
"loss": 2.126,
"step": 148
},
{
"epoch": 0.35434007134363854,
"grad_norm": 1.0236205683546673,
"learning_rate": 1.9782801602268306e-05,
"loss": 1.9399,
"step": 149
},
{
"epoch": 0.356718192627824,
"grad_norm": 0.983049547537296,
"learning_rate": 1.9778812853502592e-05,
"loss": 2.0336,
"step": 150
},
{
"epoch": 0.3590963139120095,
"grad_norm": 1.0856474713800959,
"learning_rate": 1.9774788220967968e-05,
"loss": 2.2103,
"step": 151
},
{
"epoch": 0.361474435196195,
"grad_norm": 1.098143269144179,
"learning_rate": 1.9770727719432994e-05,
"loss": 2.1425,
"step": 152
},
{
"epoch": 0.3638525564803805,
"grad_norm": 1.1908904777112574,
"learning_rate": 1.9766631363797852e-05,
"loss": 2.2516,
"step": 153
},
{
"epoch": 0.36623067776456597,
"grad_norm": 1.1823343263781934,
"learning_rate": 1.9762499169094288e-05,
"loss": 2.0991,
"step": 154
},
{
"epoch": 0.3686087990487515,
"grad_norm": 1.1543274307271654,
"learning_rate": 1.9758331150485576e-05,
"loss": 2.2917,
"step": 155
},
{
"epoch": 0.37098692033293695,
"grad_norm": 1.1828452156246019,
"learning_rate": 1.9754127323266426e-05,
"loss": 2.3577,
"step": 156
},
{
"epoch": 0.3733650416171225,
"grad_norm": 1.2458434785978698,
"learning_rate": 1.9749887702862972e-05,
"loss": 2.2291,
"step": 157
},
{
"epoch": 0.375743162901308,
"grad_norm": 1.0632348458757013,
"learning_rate": 1.9745612304832672e-05,
"loss": 2.495,
"step": 158
},
{
"epoch": 0.37812128418549346,
"grad_norm": 1.2413557275846534,
"learning_rate": 1.9741301144864284e-05,
"loss": 2.3006,
"step": 159
},
{
"epoch": 0.380499405469679,
"grad_norm": 1.068837985332943,
"learning_rate": 1.9736954238777793e-05,
"loss": 2.2228,
"step": 160
},
{
"epoch": 0.38287752675386444,
"grad_norm": 1.181973772137545,
"learning_rate": 1.9732571602524353e-05,
"loss": 2.3419,
"step": 161
},
{
"epoch": 0.38525564803804996,
"grad_norm": 0.9361759344356807,
"learning_rate": 1.972815325218624e-05,
"loss": 2.2727,
"step": 162
},
{
"epoch": 0.3876337693222354,
"grad_norm": 1.2300672941710984,
"learning_rate": 1.9723699203976768e-05,
"loss": 2.3947,
"step": 163
},
{
"epoch": 0.39001189060642094,
"grad_norm": 0.9647921025871186,
"learning_rate": 1.9719209474240263e-05,
"loss": 1.8388,
"step": 164
},
{
"epoch": 0.3923900118906064,
"grad_norm": 1.1390311715526416,
"learning_rate": 1.971468407945198e-05,
"loss": 2.4054,
"step": 165
},
{
"epoch": 0.3947681331747919,
"grad_norm": 0.9830051867519547,
"learning_rate": 1.9710123036218044e-05,
"loss": 2.0355,
"step": 166
},
{
"epoch": 0.3971462544589774,
"grad_norm": 1.1244517585073737,
"learning_rate": 1.97055263612754e-05,
"loss": 2.0188,
"step": 167
},
{
"epoch": 0.3995243757431629,
"grad_norm": 1.0256020852263494,
"learning_rate": 1.9700894071491736e-05,
"loss": 2.0774,
"step": 168
},
{
"epoch": 0.40190249702734837,
"grad_norm": 1.011023720252716,
"learning_rate": 1.9696226183865436e-05,
"loss": 2.2592,
"step": 169
},
{
"epoch": 0.4042806183115339,
"grad_norm": 1.046975898884085,
"learning_rate": 1.969152271552552e-05,
"loss": 2.1791,
"step": 170
},
{
"epoch": 0.40665873959571935,
"grad_norm": 1.1800984480399852,
"learning_rate": 1.9686783683731557e-05,
"loss": 2.3941,
"step": 171
},
{
"epoch": 0.4090368608799049,
"grad_norm": 1.2459882622321672,
"learning_rate": 1.9682009105873633e-05,
"loss": 2.1522,
"step": 172
},
{
"epoch": 0.4114149821640904,
"grad_norm": 1.0732133381850257,
"learning_rate": 1.9677198999472257e-05,
"loss": 2.1233,
"step": 173
},
{
"epoch": 0.41379310344827586,
"grad_norm": 1.2405484917580802,
"learning_rate": 1.967235338217832e-05,
"loss": 2.3016,
"step": 174
},
{
"epoch": 0.4161712247324614,
"grad_norm": 1.0759940201219593,
"learning_rate": 1.9667472271773026e-05,
"loss": 2.2947,
"step": 175
},
{
"epoch": 0.41854934601664684,
"grad_norm": 1.2008734320661734,
"learning_rate": 1.9662555686167808e-05,
"loss": 2.2155,
"step": 176
},
{
"epoch": 0.42092746730083236,
"grad_norm": 0.9303619935178572,
"learning_rate": 1.965760364340429e-05,
"loss": 2.1234,
"step": 177
},
{
"epoch": 0.4233055885850178,
"grad_norm": 1.3884826767438652,
"learning_rate": 1.9652616161654204e-05,
"loss": 2.2539,
"step": 178
},
{
"epoch": 0.42568370986920334,
"grad_norm": 0.9947187673832885,
"learning_rate": 1.9647593259219328e-05,
"loss": 2.2052,
"step": 179
},
{
"epoch": 0.4280618311533888,
"grad_norm": 1.4655922792083054,
"learning_rate": 1.964253495453141e-05,
"loss": 2.1552,
"step": 180
},
{
"epoch": 0.4304399524375743,
"grad_norm": 1.1481294188693778,
"learning_rate": 1.963744126615212e-05,
"loss": 2.3942,
"step": 181
},
{
"epoch": 0.4328180737217598,
"grad_norm": 1.239760521409481,
"learning_rate": 1.9632312212772956e-05,
"loss": 2.3091,
"step": 182
},
{
"epoch": 0.4351961950059453,
"grad_norm": 1.0524654460411744,
"learning_rate": 1.9627147813215207e-05,
"loss": 2.302,
"step": 183
},
{
"epoch": 0.43757431629013077,
"grad_norm": 1.0231645108607732,
"learning_rate": 1.9621948086429847e-05,
"loss": 2.2334,
"step": 184
},
{
"epoch": 0.4399524375743163,
"grad_norm": 1.0600582051447691,
"learning_rate": 1.9616713051497496e-05,
"loss": 2.2044,
"step": 185
},
{
"epoch": 0.4423305588585018,
"grad_norm": 1.0861978175484295,
"learning_rate": 1.9611442727628344e-05,
"loss": 2.3267,
"step": 186
},
{
"epoch": 0.4447086801426873,
"grad_norm": 1.0122924353396487,
"learning_rate": 1.960613713416206e-05,
"loss": 2.2327,
"step": 187
},
{
"epoch": 0.4470868014268728,
"grad_norm": 1.1275635495135592,
"learning_rate": 1.9600796290567747e-05,
"loss": 2.2474,
"step": 188
},
{
"epoch": 0.44946492271105826,
"grad_norm": 1.0778906611663819,
"learning_rate": 1.9595420216443864e-05,
"loss": 2.2777,
"step": 189
},
{
"epoch": 0.4518430439952438,
"grad_norm": 1.0593499669893551,
"learning_rate": 1.9590008931518133e-05,
"loss": 2.4937,
"step": 190
},
{
"epoch": 0.45422116527942924,
"grad_norm": 1.0887914371115388,
"learning_rate": 1.9584562455647494e-05,
"loss": 2.2577,
"step": 191
},
{
"epoch": 0.45659928656361476,
"grad_norm": 1.0280779311785984,
"learning_rate": 1.9579080808818035e-05,
"loss": 2.2352,
"step": 192
},
{
"epoch": 0.4589774078478002,
"grad_norm": 1.1201705856067985,
"learning_rate": 1.9573564011144873e-05,
"loss": 2.1482,
"step": 193
},
{
"epoch": 0.46135552913198574,
"grad_norm": 1.0039435227655624,
"learning_rate": 1.9568012082872148e-05,
"loss": 2.1069,
"step": 194
},
{
"epoch": 0.4637336504161712,
"grad_norm": 1.0523831000821406,
"learning_rate": 1.9562425044372884e-05,
"loss": 1.9268,
"step": 195
},
{
"epoch": 0.4661117717003567,
"grad_norm": 1.0635880350342213,
"learning_rate": 1.9556802916148963e-05,
"loss": 2.2722,
"step": 196
},
{
"epoch": 0.4684898929845422,
"grad_norm": 2.4351848601787287,
"learning_rate": 1.955114571883102e-05,
"loss": 2.1402,
"step": 197
},
{
"epoch": 0.4708680142687277,
"grad_norm": 1.2199308274597462,
"learning_rate": 1.9545453473178384e-05,
"loss": 2.2599,
"step": 198
},
{
"epoch": 0.47324613555291317,
"grad_norm": 0.9936114796299212,
"learning_rate": 1.9539726200078987e-05,
"loss": 2.0662,
"step": 199
},
{
"epoch": 0.4756242568370987,
"grad_norm": 1.0692703333507547,
"learning_rate": 1.9533963920549307e-05,
"loss": 2.3739,
"step": 200
},
{
"epoch": 0.4780023781212842,
"grad_norm": 1.0406002686664542,
"learning_rate": 1.9528166655734267e-05,
"loss": 2.3611,
"step": 201
},
{
"epoch": 0.4803804994054697,
"grad_norm": 1.9375905536343168,
"learning_rate": 1.9522334426907185e-05,
"loss": 2.0971,
"step": 202
},
{
"epoch": 0.4827586206896552,
"grad_norm": 1.024548704059581,
"learning_rate": 1.951646725546966e-05,
"loss": 2.2498,
"step": 203
},
{
"epoch": 0.48513674197384066,
"grad_norm": 1.0033895284405978,
"learning_rate": 1.9510565162951538e-05,
"loss": 2.299,
"step": 204
},
{
"epoch": 0.4875148632580262,
"grad_norm": 1.541631519071697,
"learning_rate": 1.950462817101079e-05,
"loss": 2.4076,
"step": 205
},
{
"epoch": 0.48989298454221164,
"grad_norm": 0.9499702987331401,
"learning_rate": 1.9498656301433466e-05,
"loss": 2.0754,
"step": 206
},
{
"epoch": 0.49227110582639716,
"grad_norm": 1.099383371761328,
"learning_rate": 1.9492649576133594e-05,
"loss": 2.2514,
"step": 207
},
{
"epoch": 0.4946492271105826,
"grad_norm": 0.9296431838496088,
"learning_rate": 1.94866080171531e-05,
"loss": 2.2308,
"step": 208
},
{
"epoch": 0.49702734839476814,
"grad_norm": 4.140796209905845,
"learning_rate": 1.9480531646661753e-05,
"loss": 2.4388,
"step": 209
},
{
"epoch": 0.4994054696789536,
"grad_norm": 1.011142238194789,
"learning_rate": 1.9474420486957045e-05,
"loss": 2.2414,
"step": 210
},
{
"epoch": 0.4994054696789536,
"eval_loss": 2.49302339553833,
"eval_runtime": 65.8636,
"eval_samples_per_second": 9.702,
"eval_steps_per_second": 1.215,
"step": 210
},
{
"epoch": 0.5017835909631391,
"grad_norm": 0.9941401750604694,
"learning_rate": 1.9468274560464134e-05,
"loss": 2.2182,
"step": 211
},
{
"epoch": 0.5041617122473246,
"grad_norm": 1.035068690961865,
"learning_rate": 1.9462093889735766e-05,
"loss": 2.3569,
"step": 212
},
{
"epoch": 0.5065398335315101,
"grad_norm": 1.5052185888965133,
"learning_rate": 1.945587849745217e-05,
"loss": 2.4474,
"step": 213
},
{
"epoch": 0.5089179548156956,
"grad_norm": 1.01730612362564,
"learning_rate": 1.944962840642099e-05,
"loss": 2.3912,
"step": 214
},
{
"epoch": 0.5112960760998811,
"grad_norm": 0.9513743221196854,
"learning_rate": 1.9443343639577206e-05,
"loss": 2.2842,
"step": 215
},
{
"epoch": 0.5136741973840666,
"grad_norm": 0.9759286606572132,
"learning_rate": 1.943702421998303e-05,
"loss": 2.3117,
"step": 216
},
{
"epoch": 0.5160523186682521,
"grad_norm": 0.9357816333722543,
"learning_rate": 1.9430670170827844e-05,
"loss": 2.1091,
"step": 217
},
{
"epoch": 0.5184304399524375,
"grad_norm": 1.1804547022456764,
"learning_rate": 1.94242815154281e-05,
"loss": 2.2333,
"step": 218
},
{
"epoch": 0.5208085612366231,
"grad_norm": 1.0289463651234612,
"learning_rate": 1.9417858277227244e-05,
"loss": 2.0495,
"step": 219
},
{
"epoch": 0.5231866825208086,
"grad_norm": 0.9297621773216459,
"learning_rate": 1.9411400479795618e-05,
"loss": 2.2102,
"step": 220
},
{
"epoch": 0.525564803804994,
"grad_norm": 0.9468558483075745,
"learning_rate": 1.9404908146830383e-05,
"loss": 2.2593,
"step": 221
},
{
"epoch": 0.5279429250891795,
"grad_norm": 1.03418169808611,
"learning_rate": 1.9398381302155435e-05,
"loss": 2.185,
"step": 222
},
{
"epoch": 0.5303210463733651,
"grad_norm": 0.84914434823741,
"learning_rate": 1.93918199697213e-05,
"loss": 1.8499,
"step": 223
},
{
"epoch": 0.5326991676575505,
"grad_norm": 1.0824133835143805,
"learning_rate": 1.9385224173605072e-05,
"loss": 2.2171,
"step": 224
},
{
"epoch": 0.535077288941736,
"grad_norm": 0.9478237708740055,
"learning_rate": 1.9378593938010302e-05,
"loss": 2.2397,
"step": 225
},
{
"epoch": 0.5374554102259215,
"grad_norm": 2.4277055664527065,
"learning_rate": 1.937192928726692e-05,
"loss": 2.2262,
"step": 226
},
{
"epoch": 0.539833531510107,
"grad_norm": 1.5223794932859396,
"learning_rate": 1.936523024583115e-05,
"loss": 2.2664,
"step": 227
},
{
"epoch": 0.5422116527942925,
"grad_norm": 1.2991334808397113,
"learning_rate": 1.9358496838285408e-05,
"loss": 2.1967,
"step": 228
},
{
"epoch": 0.544589774078478,
"grad_norm": 1.132823106642245,
"learning_rate": 1.9351729089338214e-05,
"loss": 2.2516,
"step": 229
},
{
"epoch": 0.5469678953626635,
"grad_norm": 1.0872118899678849,
"learning_rate": 1.9344927023824112e-05,
"loss": 2.3602,
"step": 230
},
{
"epoch": 0.549346016646849,
"grad_norm": 0.9596136441748024,
"learning_rate": 1.933809066670357e-05,
"loss": 2.3944,
"step": 231
},
{
"epoch": 0.5517241379310345,
"grad_norm": 1.3340981902899025,
"learning_rate": 1.9331220043062894e-05,
"loss": 1.8085,
"step": 232
},
{
"epoch": 0.5541022592152199,
"grad_norm": 0.9780209403031624,
"learning_rate": 1.9324315178114127e-05,
"loss": 2.2626,
"step": 233
},
{
"epoch": 0.5564803804994055,
"grad_norm": 0.8441113781121591,
"learning_rate": 1.9317376097194964e-05,
"loss": 1.904,
"step": 234
},
{
"epoch": 0.558858501783591,
"grad_norm": 0.9000708142025062,
"learning_rate": 1.9310402825768655e-05,
"loss": 2.1239,
"step": 235
},
{
"epoch": 0.5612366230677764,
"grad_norm": 0.9819927724537073,
"learning_rate": 1.9303395389423918e-05,
"loss": 2.2526,
"step": 236
},
{
"epoch": 0.5636147443519619,
"grad_norm": 0.9835396213176555,
"learning_rate": 1.9296353813874838e-05,
"loss": 2.1293,
"step": 237
},
{
"epoch": 0.5659928656361475,
"grad_norm": 0.8705520689865173,
"learning_rate": 1.9289278124960777e-05,
"loss": 1.9911,
"step": 238
},
{
"epoch": 0.5683709869203329,
"grad_norm": 1.0080028925541162,
"learning_rate": 1.9282168348646268e-05,
"loss": 2.0903,
"step": 239
},
{
"epoch": 0.5707491082045184,
"grad_norm": 1.0304202570689127,
"learning_rate": 1.927502451102095e-05,
"loss": 2.0784,
"step": 240
},
{
"epoch": 0.5731272294887039,
"grad_norm": 1.0836386452615874,
"learning_rate": 1.926784663829943e-05,
"loss": 2.2297,
"step": 241
},
{
"epoch": 0.5755053507728894,
"grad_norm": 0.972327672270083,
"learning_rate": 1.926063475682121e-05,
"loss": 2.2518,
"step": 242
},
{
"epoch": 0.5778834720570749,
"grad_norm": 1.1250553185490555,
"learning_rate": 1.9253388893050612e-05,
"loss": 2.2419,
"step": 243
},
{
"epoch": 0.5802615933412604,
"grad_norm": 1.6628057931903235,
"learning_rate": 1.924610907357663e-05,
"loss": 2.2409,
"step": 244
},
{
"epoch": 0.582639714625446,
"grad_norm": 1.3164993193299928,
"learning_rate": 1.9238795325112867e-05,
"loss": 2.4277,
"step": 245
},
{
"epoch": 0.5850178359096314,
"grad_norm": 0.9607216662774912,
"learning_rate": 1.9231447674497444e-05,
"loss": 2.0357,
"step": 246
},
{
"epoch": 0.5873959571938169,
"grad_norm": 0.9753564233974246,
"learning_rate": 1.922406614869287e-05,
"loss": 2.1743,
"step": 247
},
{
"epoch": 0.5897740784780023,
"grad_norm": 0.9514731158004311,
"learning_rate": 1.9216650774785975e-05,
"loss": 2.2775,
"step": 248
},
{
"epoch": 0.5921521997621879,
"grad_norm": 1.1514073630104866,
"learning_rate": 1.9209201579987777e-05,
"loss": 2.0307,
"step": 249
},
{
"epoch": 0.5945303210463734,
"grad_norm": 1.0070268666833808,
"learning_rate": 1.9201718591633417e-05,
"loss": 2.405,
"step": 250
},
{
"epoch": 0.5969084423305588,
"grad_norm": 0.9724192667298419,
"learning_rate": 1.9194201837182045e-05,
"loss": 2.3375,
"step": 251
},
{
"epoch": 0.5992865636147443,
"grad_norm": 0.9184225452723915,
"learning_rate": 1.9186651344216703e-05,
"loss": 2.0576,
"step": 252
},
{
"epoch": 0.6016646848989299,
"grad_norm": 0.9277770597042948,
"learning_rate": 1.9179067140444246e-05,
"loss": 2.065,
"step": 253
},
{
"epoch": 0.6040428061831153,
"grad_norm": 0.9971241617303254,
"learning_rate": 1.9171449253695233e-05,
"loss": 2.0388,
"step": 254
},
{
"epoch": 0.6064209274673008,
"grad_norm": 0.968526625269074,
"learning_rate": 1.9163797711923822e-05,
"loss": 2.2154,
"step": 255
},
{
"epoch": 0.6087990487514863,
"grad_norm": 1.0974259939522593,
"learning_rate": 1.9156112543207674e-05,
"loss": 2.0897,
"step": 256
},
{
"epoch": 0.6111771700356718,
"grad_norm": 1.0661979784688387,
"learning_rate": 1.9148393775747842e-05,
"loss": 2.3986,
"step": 257
},
{
"epoch": 0.6135552913198573,
"grad_norm": 1.3741439020152701,
"learning_rate": 1.9140641437868664e-05,
"loss": 2.1249,
"step": 258
},
{
"epoch": 0.6159334126040428,
"grad_norm": 0.9207355567454468,
"learning_rate": 1.913285555801768e-05,
"loss": 2.2715,
"step": 259
},
{
"epoch": 0.6183115338882283,
"grad_norm": 1.0171260467381849,
"learning_rate": 1.9125036164765502e-05,
"loss": 2.2638,
"step": 260
},
{
"epoch": 0.6206896551724138,
"grad_norm": 0.9849622761444283,
"learning_rate": 1.9117183286805726e-05,
"loss": 2.148,
"step": 261
},
{
"epoch": 0.6230677764565993,
"grad_norm": 1.384730629468189,
"learning_rate": 1.9109296952954826e-05,
"loss": 2.2442,
"step": 262
},
{
"epoch": 0.6254458977407847,
"grad_norm": 1.066860211861379,
"learning_rate": 1.9101377192152033e-05,
"loss": 2.451,
"step": 263
},
{
"epoch": 0.6278240190249703,
"grad_norm": 1.1537333721148562,
"learning_rate": 1.909342403345925e-05,
"loss": 2.3076,
"step": 264
},
{
"epoch": 0.6302021403091558,
"grad_norm": 1.388798426336895,
"learning_rate": 1.9085437506060925e-05,
"loss": 2.3458,
"step": 265
},
{
"epoch": 0.6325802615933412,
"grad_norm": 1.057136103530956,
"learning_rate": 1.9077417639263966e-05,
"loss": 2.1901,
"step": 266
},
{
"epoch": 0.6349583828775267,
"grad_norm": 1.0267855180849,
"learning_rate": 1.906936446249761e-05,
"loss": 2.2597,
"step": 267
},
{
"epoch": 0.6373365041617123,
"grad_norm": 0.984822259782258,
"learning_rate": 1.906127800531333e-05,
"loss": 1.9602,
"step": 268
},
{
"epoch": 0.6397146254458977,
"grad_norm": 1.0914462759031294,
"learning_rate": 1.905315829738473e-05,
"loss": 2.4246,
"step": 269
},
{
"epoch": 0.6420927467300832,
"grad_norm": 1.007066333419377,
"learning_rate": 1.9045005368507418e-05,
"loss": 2.2638,
"step": 270
},
{
"epoch": 0.6444708680142688,
"grad_norm": 0.8517882887076221,
"learning_rate": 1.9036819248598914e-05,
"loss": 1.9905,
"step": 271
},
{
"epoch": 0.6468489892984542,
"grad_norm": 1.0217039586834744,
"learning_rate": 1.9028599967698533e-05,
"loss": 2.2177,
"step": 272
},
{
"epoch": 0.6492271105826397,
"grad_norm": 0.9930930260750681,
"learning_rate": 1.902034755596727e-05,
"loss": 2.093,
"step": 273
},
{
"epoch": 0.6516052318668252,
"grad_norm": 1.4476621256505835,
"learning_rate": 1.9012062043687713e-05,
"loss": 2.2535,
"step": 274
},
{
"epoch": 0.6539833531510107,
"grad_norm": 0.9679973366177126,
"learning_rate": 1.9003743461263887e-05,
"loss": 2.2304,
"step": 275
},
{
"epoch": 0.6563614744351962,
"grad_norm": 1.0576995279849084,
"learning_rate": 1.899539183922119e-05,
"loss": 2.0678,
"step": 276
},
{
"epoch": 0.6587395957193817,
"grad_norm": 0.9196543804308106,
"learning_rate": 1.8987007208206254e-05,
"loss": 2.1106,
"step": 277
},
{
"epoch": 0.6611177170035671,
"grad_norm": 1.0412924135468502,
"learning_rate": 1.8978589598986838e-05,
"loss": 2.3659,
"step": 278
},
{
"epoch": 0.6634958382877527,
"grad_norm": 1.1221322457101093,
"learning_rate": 1.8970139042451712e-05,
"loss": 2.0853,
"step": 279
},
{
"epoch": 0.6658739595719382,
"grad_norm": 0.8927332524290843,
"learning_rate": 1.8961655569610557e-05,
"loss": 2.2388,
"step": 280
},
{
"epoch": 0.6682520808561236,
"grad_norm": 1.0304293869359518,
"learning_rate": 1.8953139211593838e-05,
"loss": 2.381,
"step": 281
},
{
"epoch": 0.6706302021403091,
"grad_norm": 1.0508469543032843,
"learning_rate": 1.8944589999652687e-05,
"loss": 2.214,
"step": 282
},
{
"epoch": 0.6730083234244947,
"grad_norm": 0.8545452252368634,
"learning_rate": 1.8936007965158806e-05,
"loss": 2.0308,
"step": 283
},
{
"epoch": 0.6753864447086801,
"grad_norm": 1.3101079437708238,
"learning_rate": 1.8927393139604327e-05,
"loss": 2.3389,
"step": 284
},
{
"epoch": 0.6777645659928656,
"grad_norm": 0.8257605335721794,
"learning_rate": 1.8918745554601726e-05,
"loss": 1.9511,
"step": 285
},
{
"epoch": 0.6801426872770512,
"grad_norm": 0.9918391764529991,
"learning_rate": 1.891006524188368e-05,
"loss": 2.3694,
"step": 286
},
{
"epoch": 0.6825208085612366,
"grad_norm": 0.8841124075255721,
"learning_rate": 1.8901352233302964e-05,
"loss": 2.1977,
"step": 287
},
{
"epoch": 0.6848989298454221,
"grad_norm": 0.8100832981740426,
"learning_rate": 1.8892606560832335e-05,
"loss": 1.9161,
"step": 288
},
{
"epoch": 0.6872770511296076,
"grad_norm": 0.8738803349826288,
"learning_rate": 1.8883828256564413e-05,
"loss": 2.2419,
"step": 289
},
{
"epoch": 0.6896551724137931,
"grad_norm": 0.8961748350433363,
"learning_rate": 1.8875017352711547e-05,
"loss": 2.1639,
"step": 290
},
{
"epoch": 0.6920332936979786,
"grad_norm": 1.0902732587802257,
"learning_rate": 1.886617388160573e-05,
"loss": 2.4486,
"step": 291
},
{
"epoch": 0.6944114149821641,
"grad_norm": 0.9653296144522788,
"learning_rate": 1.8857297875698455e-05,
"loss": 2.0904,
"step": 292
},
{
"epoch": 0.6967895362663495,
"grad_norm": 1.0898489125890396,
"learning_rate": 1.8848389367560604e-05,
"loss": 2.2726,
"step": 293
},
{
"epoch": 0.6991676575505351,
"grad_norm": 0.8892982592369411,
"learning_rate": 1.883944838988232e-05,
"loss": 1.8982,
"step": 294
},
{
"epoch": 0.7015457788347206,
"grad_norm": 0.9481426246994946,
"learning_rate": 1.8830474975472904e-05,
"loss": 2.1466,
"step": 295
},
{
"epoch": 0.703923900118906,
"grad_norm": 0.8426682151678768,
"learning_rate": 1.8821469157260687e-05,
"loss": 2.1086,
"step": 296
},
{
"epoch": 0.7063020214030915,
"grad_norm": 1.0388273479050054,
"learning_rate": 1.8812430968292888e-05,
"loss": 2.3994,
"step": 297
},
{
"epoch": 0.7086801426872771,
"grad_norm": 0.9544054737799986,
"learning_rate": 1.8803360441735533e-05,
"loss": 2.0918,
"step": 298
},
{
"epoch": 0.7110582639714625,
"grad_norm": 0.9333443425947632,
"learning_rate": 1.8794257610873307e-05,
"loss": 2.2388,
"step": 299
},
{
"epoch": 0.713436385255648,
"grad_norm": 0.9411120276164652,
"learning_rate": 1.8785122509109425e-05,
"loss": 2.1787,
"step": 300
},
{
"epoch": 0.7158145065398336,
"grad_norm": 0.899942759501507,
"learning_rate": 1.877595516996554e-05,
"loss": 2.2007,
"step": 301
},
{
"epoch": 0.718192627824019,
"grad_norm": 0.952238040187914,
"learning_rate": 1.8766755627081586e-05,
"loss": 2.1491,
"step": 302
},
{
"epoch": 0.7205707491082045,
"grad_norm": 1.0704167928673332,
"learning_rate": 1.875752391421568e-05,
"loss": 2.4306,
"step": 303
},
{
"epoch": 0.72294887039239,
"grad_norm": 1.0126343580756334,
"learning_rate": 1.8748260065243985e-05,
"loss": 2.3258,
"step": 304
},
{
"epoch": 0.7253269916765755,
"grad_norm": 0.9203276329051738,
"learning_rate": 1.8738964114160586e-05,
"loss": 2.0469,
"step": 305
},
{
"epoch": 0.727705112960761,
"grad_norm": 1.0298534605384366,
"learning_rate": 1.8729636095077368e-05,
"loss": 2.3807,
"step": 306
},
{
"epoch": 0.7300832342449465,
"grad_norm": 0.9102516225979571,
"learning_rate": 1.8720276042223896e-05,
"loss": 2.0705,
"step": 307
},
{
"epoch": 0.7324613555291319,
"grad_norm": 0.8450936687346169,
"learning_rate": 1.8710883989947278e-05,
"loss": 1.9468,
"step": 308
},
{
"epoch": 0.7348394768133175,
"grad_norm": 10.37449210680128,
"learning_rate": 1.870145997271206e-05,
"loss": 2.2267,
"step": 309
},
{
"epoch": 0.737217598097503,
"grad_norm": 0.9380258905092027,
"learning_rate": 1.8692004025100054e-05,
"loss": 2.2396,
"step": 310
},
{
"epoch": 0.7395957193816884,
"grad_norm": 0.9949249727532327,
"learning_rate": 1.868251618181028e-05,
"loss": 2.2737,
"step": 311
},
{
"epoch": 0.7419738406658739,
"grad_norm": 0.8650417909396523,
"learning_rate": 1.8672996477658767e-05,
"loss": 2.2057,
"step": 312
},
{
"epoch": 0.7443519619500595,
"grad_norm": 0.9963224791105781,
"learning_rate": 1.866344494757848e-05,
"loss": 2.3958,
"step": 313
},
{
"epoch": 0.746730083234245,
"grad_norm": 0.9782743745256545,
"learning_rate": 1.8653861626619166e-05,
"loss": 2.105,
"step": 314
},
{
"epoch": 0.7491082045184304,
"grad_norm": 0.9991591823799537,
"learning_rate": 1.8644246549947226e-05,
"loss": 2.3339,
"step": 315
},
{
"epoch": 0.7491082045184304,
"eval_loss": 2.453089952468872,
"eval_runtime": 65.8189,
"eval_samples_per_second": 9.708,
"eval_steps_per_second": 1.215,
"step": 315
},
{
"epoch": 0.751486325802616,
"grad_norm": 0.8595275246052216,
"learning_rate": 1.8634599752845594e-05,
"loss": 2.2405,
"step": 316
},
{
"epoch": 0.7538644470868014,
"grad_norm": 0.9508091435584168,
"learning_rate": 1.86249212707136e-05,
"loss": 2.0259,
"step": 317
},
{
"epoch": 0.7562425683709869,
"grad_norm": 1.007143450486727,
"learning_rate": 1.861521113906684e-05,
"loss": 2.2581,
"step": 318
},
{
"epoch": 0.7586206896551724,
"grad_norm": 0.8694388672608973,
"learning_rate": 1.8605469393537062e-05,
"loss": 2.1028,
"step": 319
},
{
"epoch": 0.760998810939358,
"grad_norm": 1.1523113038633086,
"learning_rate": 1.8595696069872013e-05,
"loss": 2.5605,
"step": 320
},
{
"epoch": 0.7633769322235434,
"grad_norm": 1.0752431273595597,
"learning_rate": 1.8585891203935315e-05,
"loss": 2.179,
"step": 321
},
{
"epoch": 0.7657550535077289,
"grad_norm": 1.4087896659381576,
"learning_rate": 1.8576054831706348e-05,
"loss": 2.3727,
"step": 322
},
{
"epoch": 0.7681331747919143,
"grad_norm": 1.0235487680455349,
"learning_rate": 1.856618698928009e-05,
"loss": 2.0918,
"step": 323
},
{
"epoch": 0.7705112960760999,
"grad_norm": 0.9904332707712578,
"learning_rate": 1.8556287712867006e-05,
"loss": 2.2627,
"step": 324
},
{
"epoch": 0.7728894173602854,
"grad_norm": 0.9900875612198691,
"learning_rate": 1.8546357038792918e-05,
"loss": 2.2281,
"step": 325
},
{
"epoch": 0.7752675386444708,
"grad_norm": 0.8882946865776852,
"learning_rate": 1.8536395003498857e-05,
"loss": 2.2966,
"step": 326
},
{
"epoch": 0.7776456599286563,
"grad_norm": 0.9917259649244573,
"learning_rate": 1.8526401643540924e-05,
"loss": 2.2903,
"step": 327
},
{
"epoch": 0.7800237812128419,
"grad_norm": 1.0110335104954997,
"learning_rate": 1.8516376995590185e-05,
"loss": 2.2732,
"step": 328
},
{
"epoch": 0.7824019024970273,
"grad_norm": 1.0613748791083453,
"learning_rate": 1.8506321096432516e-05,
"loss": 2.1801,
"step": 329
},
{
"epoch": 0.7847800237812128,
"grad_norm": 0.8280074594268549,
"learning_rate": 1.849623398296846e-05,
"loss": 2.0413,
"step": 330
},
{
"epoch": 0.7871581450653984,
"grad_norm": 1.061771296332105,
"learning_rate": 1.84861156922131e-05,
"loss": 2.3187,
"step": 331
},
{
"epoch": 0.7895362663495838,
"grad_norm": 1.003700059201312,
"learning_rate": 1.8475966261295947e-05,
"loss": 2.2117,
"step": 332
},
{
"epoch": 0.7919143876337693,
"grad_norm": 1.0346094081558943,
"learning_rate": 1.8465785727460763e-05,
"loss": 2.35,
"step": 333
},
{
"epoch": 0.7942925089179548,
"grad_norm": 1.0628441168006169,
"learning_rate": 1.845557412806545e-05,
"loss": 2.2907,
"step": 334
},
{
"epoch": 0.7966706302021404,
"grad_norm": 0.9035239279839531,
"learning_rate": 1.8445331500581905e-05,
"loss": 2.3114,
"step": 335
},
{
"epoch": 0.7990487514863258,
"grad_norm": 0.8520631758698989,
"learning_rate": 1.8435057882595885e-05,
"loss": 2.236,
"step": 336
},
{
"epoch": 0.8014268727705113,
"grad_norm": 0.9909579691053704,
"learning_rate": 1.8424753311806867e-05,
"loss": 2.164,
"step": 337
},
{
"epoch": 0.8038049940546967,
"grad_norm": 1.3497517322085937,
"learning_rate": 1.8414417826027907e-05,
"loss": 2.1371,
"step": 338
},
{
"epoch": 0.8061831153388823,
"grad_norm": 0.9733174734597209,
"learning_rate": 1.840405146318552e-05,
"loss": 2.2605,
"step": 339
},
{
"epoch": 0.8085612366230678,
"grad_norm": 0.991877145495908,
"learning_rate": 1.8393654261319504e-05,
"loss": 2.2597,
"step": 340
},
{
"epoch": 0.8109393579072532,
"grad_norm": 0.8567918308119495,
"learning_rate": 1.8383226258582833e-05,
"loss": 2.0034,
"step": 341
},
{
"epoch": 0.8133174791914387,
"grad_norm": 1.065116900935523,
"learning_rate": 1.837276749324151e-05,
"loss": 2.3297,
"step": 342
},
{
"epoch": 0.8156956004756243,
"grad_norm": 0.945929516741479,
"learning_rate": 1.8362278003674417e-05,
"loss": 2.1654,
"step": 343
},
{
"epoch": 0.8180737217598097,
"grad_norm": 1.2321854339549771,
"learning_rate": 1.8351757828373183e-05,
"loss": 2.3218,
"step": 344
},
{
"epoch": 0.8204518430439952,
"grad_norm": 0.8769538008656999,
"learning_rate": 1.8341207005942033e-05,
"loss": 1.9733,
"step": 345
},
{
"epoch": 0.8228299643281808,
"grad_norm": 0.9563552488015744,
"learning_rate": 1.8330625575097663e-05,
"loss": 1.8932,
"step": 346
},
{
"epoch": 0.8252080856123662,
"grad_norm": 0.9405864126962645,
"learning_rate": 1.8320013574669083e-05,
"loss": 2.2498,
"step": 347
},
{
"epoch": 0.8275862068965517,
"grad_norm": 0.910096309278826,
"learning_rate": 1.8309371043597472e-05,
"loss": 2.0247,
"step": 348
},
{
"epoch": 0.8299643281807372,
"grad_norm": 0.8784633584470368,
"learning_rate": 1.829869802093606e-05,
"loss": 1.9894,
"step": 349
},
{
"epoch": 0.8323424494649228,
"grad_norm": 0.8541120126588495,
"learning_rate": 1.8287994545849948e-05,
"loss": 2.0511,
"step": 350
},
{
"epoch": 0.8347205707491082,
"grad_norm": 0.9979547733824415,
"learning_rate": 1.8277260657615993e-05,
"loss": 2.0893,
"step": 351
},
{
"epoch": 0.8370986920332937,
"grad_norm": 1.0836986822911394,
"learning_rate": 1.826649639562266e-05,
"loss": 2.3076,
"step": 352
},
{
"epoch": 0.8394768133174791,
"grad_norm": 1.035582585906488,
"learning_rate": 1.825570179936986e-05,
"loss": 2.2928,
"step": 353
},
{
"epoch": 0.8418549346016647,
"grad_norm": 0.8458996980228227,
"learning_rate": 1.8244876908468826e-05,
"loss": 2.0421,
"step": 354
},
{
"epoch": 0.8442330558858502,
"grad_norm": 0.9377520676275963,
"learning_rate": 1.8234021762641946e-05,
"loss": 2.2872,
"step": 355
},
{
"epoch": 0.8466111771700356,
"grad_norm": 0.9662487818099018,
"learning_rate": 1.8223136401722648e-05,
"loss": 2.2357,
"step": 356
},
{
"epoch": 0.8489892984542212,
"grad_norm": 0.9373837869014561,
"learning_rate": 1.8212220865655224e-05,
"loss": 2.3807,
"step": 357
},
{
"epoch": 0.8513674197384067,
"grad_norm": 0.8767072116092074,
"learning_rate": 1.8201275194494695e-05,
"loss": 2.0453,
"step": 358
},
{
"epoch": 0.8537455410225921,
"grad_norm": 0.9164578385360742,
"learning_rate": 1.8190299428406667e-05,
"loss": 2.1177,
"step": 359
},
{
"epoch": 0.8561236623067776,
"grad_norm": 2.207352437074081,
"learning_rate": 1.8179293607667177e-05,
"loss": 2.2001,
"step": 360
},
{
"epoch": 0.8585017835909632,
"grad_norm": 0.983155025111798,
"learning_rate": 1.8168257772662556e-05,
"loss": 2.3009,
"step": 361
},
{
"epoch": 0.8608799048751486,
"grad_norm": 0.9734880504382539,
"learning_rate": 1.8157191963889265e-05,
"loss": 2.3093,
"step": 362
},
{
"epoch": 0.8632580261593341,
"grad_norm": 0.8887247626382502,
"learning_rate": 1.8146096221953767e-05,
"loss": 2.1673,
"step": 363
},
{
"epoch": 0.8656361474435196,
"grad_norm": 0.9468700496948975,
"learning_rate": 1.8134970587572345e-05,
"loss": 2.2193,
"step": 364
},
{
"epoch": 0.8680142687277052,
"grad_norm": 0.9774148680996411,
"learning_rate": 1.8123815101570996e-05,
"loss": 2.2185,
"step": 365
},
{
"epoch": 0.8703923900118906,
"grad_norm": 1.0377594097114105,
"learning_rate": 1.8112629804885248e-05,
"loss": 2.1385,
"step": 366
},
{
"epoch": 0.8727705112960761,
"grad_norm": 0.9257473414011718,
"learning_rate": 1.8101414738560018e-05,
"loss": 2.309,
"step": 367
},
{
"epoch": 0.8751486325802615,
"grad_norm": 0.8751605326561603,
"learning_rate": 1.8090169943749477e-05,
"loss": 2.0981,
"step": 368
},
{
"epoch": 0.8775267538644471,
"grad_norm": 0.8318450854470809,
"learning_rate": 1.8078895461716867e-05,
"loss": 2.0499,
"step": 369
},
{
"epoch": 0.8799048751486326,
"grad_norm": 0.9229990617872996,
"learning_rate": 1.8067591333834382e-05,
"loss": 2.0931,
"step": 370
},
{
"epoch": 0.882282996432818,
"grad_norm": 0.8885903764744792,
"learning_rate": 1.8056257601583004e-05,
"loss": 2.1094,
"step": 371
},
{
"epoch": 0.8846611177170036,
"grad_norm": 0.8862767802235494,
"learning_rate": 1.8044894306552338e-05,
"loss": 2.1633,
"step": 372
},
{
"epoch": 0.8870392390011891,
"grad_norm": 0.8644263027922633,
"learning_rate": 1.8033501490440478e-05,
"loss": 2.1869,
"step": 373
},
{
"epoch": 0.8894173602853745,
"grad_norm": 0.8908757391357947,
"learning_rate": 1.802207919505385e-05,
"loss": 2.0867,
"step": 374
},
{
"epoch": 0.89179548156956,
"grad_norm": 0.9706439808339724,
"learning_rate": 1.801062746230705e-05,
"loss": 2.2817,
"step": 375
},
{
"epoch": 0.8941736028537456,
"grad_norm": 0.9665648406193532,
"learning_rate": 1.79991463342227e-05,
"loss": 2.109,
"step": 376
},
{
"epoch": 0.896551724137931,
"grad_norm": 0.8628669010410803,
"learning_rate": 1.798763585293128e-05,
"loss": 2.0049,
"step": 377
},
{
"epoch": 0.8989298454221165,
"grad_norm": 0.9905082003227874,
"learning_rate": 1.7976096060671e-05,
"loss": 2.3004,
"step": 378
},
{
"epoch": 0.901307966706302,
"grad_norm": 0.9857769539636477,
"learning_rate": 1.7964526999787606e-05,
"loss": 2.2076,
"step": 379
},
{
"epoch": 0.9036860879904876,
"grad_norm": 0.9874365687027726,
"learning_rate": 1.7952928712734266e-05,
"loss": 2.2918,
"step": 380
},
{
"epoch": 0.906064209274673,
"grad_norm": 0.8854059768194814,
"learning_rate": 1.7941301242071384e-05,
"loss": 2.1416,
"step": 381
},
{
"epoch": 0.9084423305588585,
"grad_norm": 0.8891010909161526,
"learning_rate": 1.792964463046646e-05,
"loss": 2.2335,
"step": 382
},
{
"epoch": 0.9108204518430439,
"grad_norm": 0.8836139495461457,
"learning_rate": 1.7917958920693923e-05,
"loss": 2.0156,
"step": 383
},
{
"epoch": 0.9131985731272295,
"grad_norm": 0.862309723766035,
"learning_rate": 1.790624415563498e-05,
"loss": 2.2668,
"step": 384
},
{
"epoch": 0.915576694411415,
"grad_norm": 0.8736380448815635,
"learning_rate": 1.7894500378277463e-05,
"loss": 2.1338,
"step": 385
},
{
"epoch": 0.9179548156956004,
"grad_norm": 0.8450572821559386,
"learning_rate": 1.7882727631715655e-05,
"loss": 2.0142,
"step": 386
},
{
"epoch": 0.920332936979786,
"grad_norm": 0.9862096866699867,
"learning_rate": 1.7870925959150155e-05,
"loss": 1.9915,
"step": 387
},
{
"epoch": 0.9227110582639715,
"grad_norm": 1.0092700839816597,
"learning_rate": 1.7859095403887697e-05,
"loss": 2.0294,
"step": 388
},
{
"epoch": 0.925089179548157,
"grad_norm": 1.1074789492464077,
"learning_rate": 1.7847236009341007e-05,
"loss": 2.2817,
"step": 389
},
{
"epoch": 0.9274673008323424,
"grad_norm": 1.1905201839698938,
"learning_rate": 1.7835347819028642e-05,
"loss": 2.1793,
"step": 390
},
{
"epoch": 0.929845422116528,
"grad_norm": 0.9783879172577963,
"learning_rate": 1.7823430876574815e-05,
"loss": 2.0936,
"step": 391
},
{
"epoch": 0.9322235434007135,
"grad_norm": 0.8747590280954227,
"learning_rate": 1.7811485225709255e-05,
"loss": 2.0459,
"step": 392
},
{
"epoch": 0.9346016646848989,
"grad_norm": 0.9379273246769285,
"learning_rate": 1.7799510910267032e-05,
"loss": 2.3024,
"step": 393
},
{
"epoch": 0.9369797859690844,
"grad_norm": 1.0154785871015584,
"learning_rate": 1.778750797418841e-05,
"loss": 2.2336,
"step": 394
},
{
"epoch": 0.93935790725327,
"grad_norm": 2.334020052582292,
"learning_rate": 1.7775476461518668e-05,
"loss": 2.2146,
"step": 395
},
{
"epoch": 0.9417360285374554,
"grad_norm": 1.028424946167679,
"learning_rate": 1.7763416416407953e-05,
"loss": 2.148,
"step": 396
},
{
"epoch": 0.9441141498216409,
"grad_norm": 0.9939249719466411,
"learning_rate": 1.7751327883111117e-05,
"loss": 2.3384,
"step": 397
},
{
"epoch": 0.9464922711058263,
"grad_norm": 0.9622526498121815,
"learning_rate": 1.773921090598754e-05,
"loss": 2.2333,
"step": 398
},
{
"epoch": 0.9488703923900119,
"grad_norm": 0.9627114271614428,
"learning_rate": 1.7727065529500986e-05,
"loss": 2.0078,
"step": 399
},
{
"epoch": 0.9512485136741974,
"grad_norm": 1.0739248368345071,
"learning_rate": 1.7714891798219432e-05,
"loss": 2.431,
"step": 400
},
{
"epoch": 0.9536266349583828,
"grad_norm": 0.9459448330629742,
"learning_rate": 1.7702689756814898e-05,
"loss": 2.195,
"step": 401
},
{
"epoch": 0.9560047562425684,
"grad_norm": 0.985098727152756,
"learning_rate": 1.7690459450063297e-05,
"loss": 2.378,
"step": 402
},
{
"epoch": 0.9583828775267539,
"grad_norm": 0.9004480308740594,
"learning_rate": 1.7678200922844256e-05,
"loss": 2.0375,
"step": 403
},
{
"epoch": 0.9607609988109393,
"grad_norm": 1.0816635115771127,
"learning_rate": 1.7665914220140964e-05,
"loss": 2.2932,
"step": 404
},
{
"epoch": 0.9631391200951248,
"grad_norm": 0.856817467594963,
"learning_rate": 1.7653599387039993e-05,
"loss": 2.1395,
"step": 405
},
{
"epoch": 0.9655172413793104,
"grad_norm": 0.9816466176158317,
"learning_rate": 1.764125646873115e-05,
"loss": 2.2032,
"step": 406
},
{
"epoch": 0.9678953626634959,
"grad_norm": 1.0159180902253635,
"learning_rate": 1.7628885510507295e-05,
"loss": 2.3814,
"step": 407
},
{
"epoch": 0.9702734839476813,
"grad_norm": 0.8765185622300491,
"learning_rate": 1.7616486557764187e-05,
"loss": 2.1377,
"step": 408
},
{
"epoch": 0.9726516052318668,
"grad_norm": 0.8581340453429611,
"learning_rate": 1.7604059656000313e-05,
"loss": 2.1687,
"step": 409
},
{
"epoch": 0.9750297265160524,
"grad_norm": 1.0106580156311586,
"learning_rate": 1.7591604850816705e-05,
"loss": 2.4389,
"step": 410
},
{
"epoch": 0.9774078478002378,
"grad_norm": 0.9102932420528898,
"learning_rate": 1.757912218791681e-05,
"loss": 2.1774,
"step": 411
},
{
"epoch": 0.9797859690844233,
"grad_norm": 0.8897815759878576,
"learning_rate": 1.7566611713106287e-05,
"loss": 2.0938,
"step": 412
},
{
"epoch": 0.9821640903686087,
"grad_norm": 0.8151179975871685,
"learning_rate": 1.7554073472292854e-05,
"loss": 2.0923,
"step": 413
},
{
"epoch": 0.9845422116527943,
"grad_norm": 0.9325421154931588,
"learning_rate": 1.7541507511486114e-05,
"loss": 2.2798,
"step": 414
},
{
"epoch": 0.9869203329369798,
"grad_norm": 0.9169184208068022,
"learning_rate": 1.75289138767974e-05,
"loss": 2.2266,
"step": 415
},
{
"epoch": 0.9892984542211652,
"grad_norm": 0.9539647094698775,
"learning_rate": 1.7516292614439586e-05,
"loss": 2.3459,
"step": 416
},
{
"epoch": 0.9916765755053508,
"grad_norm": 1.0372921455846975,
"learning_rate": 1.7503643770726924e-05,
"loss": 2.2152,
"step": 417
},
{
"epoch": 0.9940546967895363,
"grad_norm": 0.8903926236033542,
"learning_rate": 1.7490967392074897e-05,
"loss": 2.1959,
"step": 418
},
{
"epoch": 0.9964328180737217,
"grad_norm": 0.8491943369235393,
"learning_rate": 1.7478263525000003e-05,
"loss": 2.265,
"step": 419
},
{
"epoch": 0.9988109393579072,
"grad_norm": 0.9138223778798192,
"learning_rate": 1.7465532216119628e-05,
"loss": 2.2167,
"step": 420
},
{
"epoch": 0.9988109393579072,
"eval_loss": 2.4223339557647705,
"eval_runtime": 65.3623,
"eval_samples_per_second": 9.776,
"eval_steps_per_second": 1.224,
"step": 420
},
{
"epoch": 1.0,
"grad_norm": 0.9138223778798192,
"learning_rate": 1.7452773512151847e-05,
"loss": 2.272,
"step": 421
},
{
"epoch": 1.0023781212841856,
"grad_norm": 1.628296924382809,
"learning_rate": 1.7439987459915265e-05,
"loss": 1.7689,
"step": 422
},
{
"epoch": 1.004756242568371,
"grad_norm": 1.0685556763602093,
"learning_rate": 1.7427174106328846e-05,
"loss": 1.952,
"step": 423
},
{
"epoch": 1.0071343638525565,
"grad_norm": 1.1222612546095956,
"learning_rate": 1.7414333498411734e-05,
"loss": 2.0925,
"step": 424
},
{
"epoch": 1.009512485136742,
"grad_norm": 1.0714061881839094,
"learning_rate": 1.7401465683283085e-05,
"loss": 1.9147,
"step": 425
},
{
"epoch": 1.0118906064209274,
"grad_norm": 1.7138310113119204,
"learning_rate": 1.7388570708161895e-05,
"loss": 2.0451,
"step": 426
},
{
"epoch": 1.014268727705113,
"grad_norm": 1.2402516993910184,
"learning_rate": 1.7375648620366817e-05,
"loss": 2.0889,
"step": 427
},
{
"epoch": 1.0166468489892984,
"grad_norm": 1.120370001804281,
"learning_rate": 1.7362699467316005e-05,
"loss": 2.0449,
"step": 428
},
{
"epoch": 1.019024970273484,
"grad_norm": 1.1942571073271016,
"learning_rate": 1.7349723296526923e-05,
"loss": 1.9913,
"step": 429
},
{
"epoch": 1.0214030915576695,
"grad_norm": 1.193074546908054,
"learning_rate": 1.7336720155616186e-05,
"loss": 2.0203,
"step": 430
},
{
"epoch": 1.0237812128418549,
"grad_norm": 0.8336179192625718,
"learning_rate": 1.732369009229937e-05,
"loss": 1.9839,
"step": 431
},
{
"epoch": 1.0261593341260404,
"grad_norm": 1.0440055245119049,
"learning_rate": 1.731063315439084e-05,
"loss": 1.9837,
"step": 432
},
{
"epoch": 1.028537455410226,
"grad_norm": 1.0017223079381312,
"learning_rate": 1.729754938980359e-05,
"loss": 2.0407,
"step": 433
},
{
"epoch": 1.0309155766944114,
"grad_norm": 1.0631773992935534,
"learning_rate": 1.7284438846549045e-05,
"loss": 2.0097,
"step": 434
},
{
"epoch": 1.033293697978597,
"grad_norm": 0.9980542007217309,
"learning_rate": 1.7271301572736904e-05,
"loss": 2.0821,
"step": 435
},
{
"epoch": 1.0356718192627823,
"grad_norm": 1.176402924330386,
"learning_rate": 1.725813761657495e-05,
"loss": 1.8594,
"step": 436
},
{
"epoch": 1.0380499405469679,
"grad_norm": 0.9752892126553369,
"learning_rate": 1.7244947026368878e-05,
"loss": 1.9673,
"step": 437
},
{
"epoch": 1.0404280618311534,
"grad_norm": 0.8689953910803522,
"learning_rate": 1.723172985052212e-05,
"loss": 1.9404,
"step": 438
},
{
"epoch": 1.0428061831153388,
"grad_norm": 1.2475813437261176,
"learning_rate": 1.7218486137535663e-05,
"loss": 2.0084,
"step": 439
},
{
"epoch": 1.0451843043995244,
"grad_norm": 1.0792429756941824,
"learning_rate": 1.720521593600787e-05,
"loss": 1.7713,
"step": 440
},
{
"epoch": 1.04756242568371,
"grad_norm": 0.9766534238863007,
"learning_rate": 1.7191919294634318e-05,
"loss": 1.9303,
"step": 441
},
{
"epoch": 1.0499405469678953,
"grad_norm": 1.0433496430768199,
"learning_rate": 1.717859626220759e-05,
"loss": 2.0188,
"step": 442
},
{
"epoch": 1.0523186682520809,
"grad_norm": 1.119386810190894,
"learning_rate": 1.7165246887617124e-05,
"loss": 1.9778,
"step": 443
},
{
"epoch": 1.0546967895362664,
"grad_norm": 0.9743730355346135,
"learning_rate": 1.715187121984901e-05,
"loss": 1.8941,
"step": 444
},
{
"epoch": 1.0570749108204518,
"grad_norm": 1.0796463722814744,
"learning_rate": 1.7138469307985832e-05,
"loss": 1.9573,
"step": 445
},
{
"epoch": 1.0594530321046374,
"grad_norm": 1.0491377417614562,
"learning_rate": 1.7125041201206478e-05,
"loss": 1.8715,
"step": 446
},
{
"epoch": 1.0618311533888227,
"grad_norm": 1.0722587649757946,
"learning_rate": 1.711158694878595e-05,
"loss": 1.892,
"step": 447
},
{
"epoch": 1.0642092746730083,
"grad_norm": 0.9251557954985203,
"learning_rate": 1.7098106600095204e-05,
"loss": 1.8922,
"step": 448
},
{
"epoch": 1.0665873959571939,
"grad_norm": 1.0363952883035685,
"learning_rate": 1.708460020460095e-05,
"loss": 1.9076,
"step": 449
},
{
"epoch": 1.0689655172413792,
"grad_norm": 1.0688439486649002,
"learning_rate": 1.7071067811865477e-05,
"loss": 1.9799,
"step": 450
},
{
"epoch": 1.0713436385255648,
"grad_norm": 1.1027957836331184,
"learning_rate": 1.7057509471546476e-05,
"loss": 1.9104,
"step": 451
},
{
"epoch": 1.0737217598097504,
"grad_norm": 1.618486595162837,
"learning_rate": 1.7043925233396855e-05,
"loss": 2.0171,
"step": 452
},
{
"epoch": 1.0760998810939357,
"grad_norm": 1.0017778693822061,
"learning_rate": 1.703031514726455e-05,
"loss": 2.0216,
"step": 453
},
{
"epoch": 1.0784780023781213,
"grad_norm": 1.0070735770943433,
"learning_rate": 1.7016679263092352e-05,
"loss": 2.0046,
"step": 454
},
{
"epoch": 1.0808561236623069,
"grad_norm": 1.1684678408376339,
"learning_rate": 1.700301763091771e-05,
"loss": 2.0023,
"step": 455
},
{
"epoch": 1.0832342449464922,
"grad_norm": 1.0525233453838523,
"learning_rate": 1.6989330300872576e-05,
"loss": 1.7791,
"step": 456
},
{
"epoch": 1.0856123662306778,
"grad_norm": 0.9255203454055275,
"learning_rate": 1.6975617323183175e-05,
"loss": 1.9353,
"step": 457
},
{
"epoch": 1.0879904875148632,
"grad_norm": 1.0726155742434451,
"learning_rate": 1.6961878748169868e-05,
"loss": 1.8358,
"step": 458
},
{
"epoch": 1.0903686087990487,
"grad_norm": 1.0856909362569107,
"learning_rate": 1.694811462624694e-05,
"loss": 2.0117,
"step": 459
},
{
"epoch": 1.0927467300832343,
"grad_norm": 1.0205381713143487,
"learning_rate": 1.6934325007922418e-05,
"loss": 1.9674,
"step": 460
},
{
"epoch": 1.0951248513674197,
"grad_norm": 1.2186466927066344,
"learning_rate": 1.6920509943797893e-05,
"loss": 2.0536,
"step": 461
},
{
"epoch": 1.0975029726516052,
"grad_norm": 1.0835683944583578,
"learning_rate": 1.690666948456833e-05,
"loss": 2.0184,
"step": 462
},
{
"epoch": 1.0998810939357908,
"grad_norm": 0.9540266090666372,
"learning_rate": 1.689280368102188e-05,
"loss": 1.7483,
"step": 463
},
{
"epoch": 1.1022592152199762,
"grad_norm": 1.1931606541151196,
"learning_rate": 1.6878912584039698e-05,
"loss": 1.8391,
"step": 464
},
{
"epoch": 1.1046373365041617,
"grad_norm": 0.9898663043482883,
"learning_rate": 1.6864996244595757e-05,
"loss": 2.0927,
"step": 465
},
{
"epoch": 1.1070154577883473,
"grad_norm": 1.1388354501174491,
"learning_rate": 1.6851054713756653e-05,
"loss": 2.1605,
"step": 466
},
{
"epoch": 1.1093935790725327,
"grad_norm": 1.0945133712632522,
"learning_rate": 1.6837088042681423e-05,
"loss": 1.9853,
"step": 467
},
{
"epoch": 1.1117717003567182,
"grad_norm": 1.1984401216196918,
"learning_rate": 1.6823096282621366e-05,
"loss": 2.0257,
"step": 468
},
{
"epoch": 1.1141498216409036,
"grad_norm": 0.8917099119377462,
"learning_rate": 1.6809079484919834e-05,
"loss": 1.8659,
"step": 469
},
{
"epoch": 1.1165279429250892,
"grad_norm": 1.0116433318578837,
"learning_rate": 1.679503770101206e-05,
"loss": 2.2598,
"step": 470
},
{
"epoch": 1.1189060642092747,
"grad_norm": 3.175396684459374,
"learning_rate": 1.6780970982424966e-05,
"loss": 1.9356,
"step": 471
},
{
"epoch": 1.12128418549346,
"grad_norm": 1.2346613722854758,
"learning_rate": 1.6766879380776983e-05,
"loss": 1.8155,
"step": 472
},
{
"epoch": 1.1236623067776457,
"grad_norm": 1.006693147566446,
"learning_rate": 1.675276294777783e-05,
"loss": 1.8518,
"step": 473
},
{
"epoch": 1.1260404280618312,
"grad_norm": 0.8894722137287147,
"learning_rate": 1.6738621735228363e-05,
"loss": 2.0505,
"step": 474
},
{
"epoch": 1.1284185493460166,
"grad_norm": 1.103577294552584,
"learning_rate": 1.672445579502036e-05,
"loss": 1.9912,
"step": 475
},
{
"epoch": 1.1307966706302022,
"grad_norm": 1.096718928499225,
"learning_rate": 1.671026517913634e-05,
"loss": 1.8938,
"step": 476
},
{
"epoch": 1.1331747919143877,
"grad_norm": 0.9817288455594879,
"learning_rate": 1.669604993964937e-05,
"loss": 1.8101,
"step": 477
},
{
"epoch": 1.135552913198573,
"grad_norm": 1.1460045828685286,
"learning_rate": 1.668181012872288e-05,
"loss": 2.2724,
"step": 478
},
{
"epoch": 1.1379310344827587,
"grad_norm": 1.1002427752644104,
"learning_rate": 1.666754579861045e-05,
"loss": 2.0842,
"step": 479
},
{
"epoch": 1.140309155766944,
"grad_norm": 1.0583766063889544,
"learning_rate": 1.6653257001655652e-05,
"loss": 2.006,
"step": 480
},
{
"epoch": 1.1426872770511296,
"grad_norm": 1.0862329799584287,
"learning_rate": 1.6638943790291838e-05,
"loss": 2.074,
"step": 481
},
{
"epoch": 1.1450653983353152,
"grad_norm": 1.215641711818309,
"learning_rate": 1.6624606217041933e-05,
"loss": 1.9628,
"step": 482
},
{
"epoch": 1.1474435196195005,
"grad_norm": 0.9889402798664048,
"learning_rate": 1.661024433451828e-05,
"loss": 2.0147,
"step": 483
},
{
"epoch": 1.149821640903686,
"grad_norm": 0.9856498299194657,
"learning_rate": 1.6595858195422414e-05,
"loss": 1.8529,
"step": 484
},
{
"epoch": 1.1521997621878715,
"grad_norm": 1.0678057323434258,
"learning_rate": 1.6581447852544877e-05,
"loss": 1.7709,
"step": 485
},
{
"epoch": 1.154577883472057,
"grad_norm": 0.8386080848257256,
"learning_rate": 1.6567013358765045e-05,
"loss": 1.9122,
"step": 486
},
{
"epoch": 1.1569560047562426,
"grad_norm": 1.0694865554443245,
"learning_rate": 1.6552554767050898e-05,
"loss": 1.8617,
"step": 487
},
{
"epoch": 1.1593341260404282,
"grad_norm": 1.9413752963807895,
"learning_rate": 1.6538072130458853e-05,
"loss": 1.9043,
"step": 488
},
{
"epoch": 1.1617122473246135,
"grad_norm": 1.04735712587821,
"learning_rate": 1.6523565502133562e-05,
"loss": 1.8987,
"step": 489
},
{
"epoch": 1.164090368608799,
"grad_norm": 0.9832837853313156,
"learning_rate": 1.6509034935307716e-05,
"loss": 2.1317,
"step": 490
},
{
"epoch": 1.1664684898929845,
"grad_norm": 1.1810675679989917,
"learning_rate": 1.6494480483301836e-05,
"loss": 1.9238,
"step": 491
},
{
"epoch": 1.16884661117717,
"grad_norm": 0.9165494247402323,
"learning_rate": 1.6479902199524116e-05,
"loss": 2.0648,
"step": 492
},
{
"epoch": 1.1712247324613556,
"grad_norm": 0.9636466082435201,
"learning_rate": 1.6465300137470178e-05,
"loss": 2.0287,
"step": 493
},
{
"epoch": 1.173602853745541,
"grad_norm": 0.9807575958399394,
"learning_rate": 1.645067435072291e-05,
"loss": 1.8887,
"step": 494
},
{
"epoch": 1.1759809750297265,
"grad_norm": 0.9586275603088916,
"learning_rate": 1.6436024892952256e-05,
"loss": 1.8363,
"step": 495
},
{
"epoch": 1.178359096313912,
"grad_norm": 0.9457623737642574,
"learning_rate": 1.6421351817915025e-05,
"loss": 2.146,
"step": 496
},
{
"epoch": 1.1807372175980975,
"grad_norm": 1.0017147042513856,
"learning_rate": 1.6406655179454694e-05,
"loss": 2.0542,
"step": 497
},
{
"epoch": 1.183115338882283,
"grad_norm": 1.9322307606799163,
"learning_rate": 1.6391935031501193e-05,
"loss": 1.9879,
"step": 498
},
{
"epoch": 1.1854934601664684,
"grad_norm": 1.084335177141969,
"learning_rate": 1.6377191428070734e-05,
"loss": 2.0715,
"step": 499
},
{
"epoch": 1.187871581450654,
"grad_norm": 0.9850991570942521,
"learning_rate": 1.63624244232656e-05,
"loss": 1.9499,
"step": 500
},
{
"epoch": 1.1902497027348395,
"grad_norm": 1.0171080592281492,
"learning_rate": 1.6347634071273932e-05,
"loss": 1.9646,
"step": 501
},
{
"epoch": 1.192627824019025,
"grad_norm": 1.0777884529528277,
"learning_rate": 1.6332820426369567e-05,
"loss": 2.0598,
"step": 502
},
{
"epoch": 1.1950059453032105,
"grad_norm": 1.020533123546286,
"learning_rate": 1.631798354291179e-05,
"loss": 1.8633,
"step": 503
},
{
"epoch": 1.197384066587396,
"grad_norm": 1.2067764605837605,
"learning_rate": 1.6303123475345182e-05,
"loss": 1.9119,
"step": 504
},
{
"epoch": 1.1997621878715814,
"grad_norm": 1.053791154347935,
"learning_rate": 1.6288240278199393e-05,
"loss": 2.1143,
"step": 505
},
{
"epoch": 1.202140309155767,
"grad_norm": 1.0193158748451816,
"learning_rate": 1.6273334006088946e-05,
"loss": 1.9422,
"step": 506
},
{
"epoch": 1.2045184304399523,
"grad_norm": 1.1993210309578368,
"learning_rate": 1.6258404713713035e-05,
"loss": 1.9505,
"step": 507
},
{
"epoch": 1.206896551724138,
"grad_norm": 0.9278344614652998,
"learning_rate": 1.624345245585534e-05,
"loss": 1.8948,
"step": 508
},
{
"epoch": 1.2092746730083235,
"grad_norm": 0.8732547686659468,
"learning_rate": 1.62284772873838e-05,
"loss": 1.8943,
"step": 509
},
{
"epoch": 1.2116527942925088,
"grad_norm": 1.1985862989626987,
"learning_rate": 1.6213479263250433e-05,
"loss": 1.8923,
"step": 510
},
{
"epoch": 1.2140309155766944,
"grad_norm": 0.9559917937082356,
"learning_rate": 1.6198458438491124e-05,
"loss": 1.9478,
"step": 511
},
{
"epoch": 1.21640903686088,
"grad_norm": 1.0369642986640548,
"learning_rate": 1.6183414868225434e-05,
"loss": 1.932,
"step": 512
},
{
"epoch": 1.2187871581450653,
"grad_norm": 0.8541777304922422,
"learning_rate": 1.616834860765637e-05,
"loss": 2.0197,
"step": 513
},
{
"epoch": 1.221165279429251,
"grad_norm": 1.1533492531760159,
"learning_rate": 1.6153259712070225e-05,
"loss": 2.0733,
"step": 514
},
{
"epoch": 1.2235434007134365,
"grad_norm": 1.1544430837733473,
"learning_rate": 1.613814823683634e-05,
"loss": 1.9616,
"step": 515
},
{
"epoch": 1.2259215219976218,
"grad_norm": 0.904111613570178,
"learning_rate": 1.6123014237406912e-05,
"loss": 2.1678,
"step": 516
},
{
"epoch": 1.2282996432818074,
"grad_norm": 1.101310926122992,
"learning_rate": 1.6107857769316798e-05,
"loss": 1.8998,
"step": 517
},
{
"epoch": 1.2306777645659928,
"grad_norm": 0.9065269772612409,
"learning_rate": 1.6092678888183298e-05,
"loss": 2.0495,
"step": 518
},
{
"epoch": 1.2330558858501783,
"grad_norm": 1.252782833936401,
"learning_rate": 1.6077477649705963e-05,
"loss": 2.1285,
"step": 519
},
{
"epoch": 1.235434007134364,
"grad_norm": 0.9769860406730624,
"learning_rate": 1.6062254109666383e-05,
"loss": 2.0681,
"step": 520
},
{
"epoch": 1.2378121284185493,
"grad_norm": 1.1110445238181776,
"learning_rate": 1.604700832392798e-05,
"loss": 1.8933,
"step": 521
},
{
"epoch": 1.2401902497027348,
"grad_norm": 1.044536874093627,
"learning_rate": 1.603174034843582e-05,
"loss": 1.932,
"step": 522
},
{
"epoch": 1.2425683709869204,
"grad_norm": 1.271022221507985,
"learning_rate": 1.601645023921638e-05,
"loss": 2.0284,
"step": 523
},
{
"epoch": 1.2449464922711058,
"grad_norm": 1.1152833273519336,
"learning_rate": 1.600113805237737e-05,
"loss": 1.9726,
"step": 524
},
{
"epoch": 1.2473246135552913,
"grad_norm": 1.2380069046576747,
"learning_rate": 1.5985803844107502e-05,
"loss": 1.8106,
"step": 525
},
{
"epoch": 1.2473246135552913,
"eval_loss": 2.4183852672576904,
"eval_runtime": 65.6332,
"eval_samples_per_second": 9.736,
"eval_steps_per_second": 1.219,
"step": 525
},
{
"epoch": 1.249702734839477,
"grad_norm": 1.4005597471855011,
"learning_rate": 1.5970447670676314e-05,
"loss": 1.9685,
"step": 526
},
{
"epoch": 1.2520808561236623,
"grad_norm": 0.9054766632630273,
"learning_rate": 1.5955069588433932e-05,
"loss": 1.9919,
"step": 527
},
{
"epoch": 1.2544589774078478,
"grad_norm": 1.1186292190334863,
"learning_rate": 1.5939669653810882e-05,
"loss": 1.8359,
"step": 528
},
{
"epoch": 1.2568370986920332,
"grad_norm": 0.8661419871842052,
"learning_rate": 1.5924247923317874e-05,
"loss": 2.0698,
"step": 529
},
{
"epoch": 1.2592152199762188,
"grad_norm": 1.3186144918707836,
"learning_rate": 1.5908804453545608e-05,
"loss": 1.8845,
"step": 530
},
{
"epoch": 1.2615933412604043,
"grad_norm": 1.3508521915293898,
"learning_rate": 1.589333930116455e-05,
"loss": 2.0008,
"step": 531
},
{
"epoch": 1.2639714625445897,
"grad_norm": 1.0426903382069417,
"learning_rate": 1.5877852522924733e-05,
"loss": 2.14,
"step": 532
},
{
"epoch": 1.2663495838287753,
"grad_norm": 1.1270060025806774,
"learning_rate": 1.5862344175655547e-05,
"loss": 1.7585,
"step": 533
},
{
"epoch": 1.2687277051129606,
"grad_norm": 0.894696959523861,
"learning_rate": 1.584681431626553e-05,
"loss": 1.7954,
"step": 534
},
{
"epoch": 1.2711058263971462,
"grad_norm": 0.8963443148345133,
"learning_rate": 1.5831263001742168e-05,
"loss": 1.8992,
"step": 535
},
{
"epoch": 1.2734839476813318,
"grad_norm": 0.9751034618814204,
"learning_rate": 1.581569028915166e-05,
"loss": 2.0839,
"step": 536
},
{
"epoch": 1.2758620689655173,
"grad_norm": 1.2846795383089968,
"learning_rate": 1.5800096235638735e-05,
"loss": 2.0306,
"step": 537
},
{
"epoch": 1.2782401902497027,
"grad_norm": 1.1344265063076209,
"learning_rate": 1.5784480898426444e-05,
"loss": 2.0588,
"step": 538
},
{
"epoch": 1.2806183115338883,
"grad_norm": 2.3160794394882807,
"learning_rate": 1.5768844334815923e-05,
"loss": 2.2461,
"step": 539
},
{
"epoch": 1.2829964328180736,
"grad_norm": 1.2198804497684061,
"learning_rate": 1.5753186602186207e-05,
"loss": 2.0165,
"step": 540
},
{
"epoch": 1.2853745541022592,
"grad_norm": 1.0333157223752394,
"learning_rate": 1.573750775799401e-05,
"loss": 1.9158,
"step": 541
},
{
"epoch": 1.2877526753864448,
"grad_norm": 1.0933945363135282,
"learning_rate": 1.572180785977352e-05,
"loss": 2.0026,
"step": 542
},
{
"epoch": 1.2901307966706301,
"grad_norm": 1.0383257582502903,
"learning_rate": 1.570608696513618e-05,
"loss": 2.1656,
"step": 543
},
{
"epoch": 1.2925089179548157,
"grad_norm": 1.238965226787101,
"learning_rate": 1.5690345131770474e-05,
"loss": 1.8335,
"step": 544
},
{
"epoch": 1.294887039239001,
"grad_norm": 0.7431372863827379,
"learning_rate": 1.5674582417441734e-05,
"loss": 2.0428,
"step": 545
},
{
"epoch": 1.2972651605231866,
"grad_norm": 1.482428714196644,
"learning_rate": 1.5658798879991905e-05,
"loss": 2.0023,
"step": 546
},
{
"epoch": 1.2996432818073722,
"grad_norm": 1.2061227422514962,
"learning_rate": 1.564299457733935e-05,
"loss": 2.0852,
"step": 547
},
{
"epoch": 1.3020214030915578,
"grad_norm": 1.2166153788730312,
"learning_rate": 1.5627169567478627e-05,
"loss": 1.7886,
"step": 548
},
{
"epoch": 1.3043995243757431,
"grad_norm": 0.9980529341355449,
"learning_rate": 1.561132390848029e-05,
"loss": 1.8238,
"step": 549
},
{
"epoch": 1.3067776456599287,
"grad_norm": 1.0516168634342369,
"learning_rate": 1.5595457658490643e-05,
"loss": 1.9991,
"step": 550
},
{
"epoch": 1.309155766944114,
"grad_norm": 1.1353824503332424,
"learning_rate": 1.5579570875731572e-05,
"loss": 2.2081,
"step": 551
},
{
"epoch": 1.3115338882282996,
"grad_norm": 1.0157342209528704,
"learning_rate": 1.55636636185003e-05,
"loss": 1.9142,
"step": 552
},
{
"epoch": 1.3139120095124852,
"grad_norm": 1.3048380581725043,
"learning_rate": 1.5547735945169188e-05,
"loss": 1.9015,
"step": 553
},
{
"epoch": 1.3162901307966706,
"grad_norm": 1.0849291353506971,
"learning_rate": 1.55317879141855e-05,
"loss": 1.892,
"step": 554
},
{
"epoch": 1.3186682520808561,
"grad_norm": 1.103994377178752,
"learning_rate": 1.5515819584071216e-05,
"loss": 2.0736,
"step": 555
},
{
"epoch": 1.3210463733650415,
"grad_norm": 0.87333743581362,
"learning_rate": 1.5499831013422804e-05,
"loss": 1.6585,
"step": 556
},
{
"epoch": 1.323424494649227,
"grad_norm": 1.5096036417812366,
"learning_rate": 1.5483822260911002e-05,
"loss": 2.0984,
"step": 557
},
{
"epoch": 1.3258026159334126,
"grad_norm": 0.9402021272644058,
"learning_rate": 1.5467793385280602e-05,
"loss": 1.919,
"step": 558
},
{
"epoch": 1.3281807372175982,
"grad_norm": 1.010147060155673,
"learning_rate": 1.5451744445350252e-05,
"loss": 1.789,
"step": 559
},
{
"epoch": 1.3305588585017836,
"grad_norm": 1.018932157068251,
"learning_rate": 1.5435675500012212e-05,
"loss": 1.9567,
"step": 560
},
{
"epoch": 1.3329369797859691,
"grad_norm": 0.9622353435965788,
"learning_rate": 1.5419586608232163e-05,
"loss": 2.102,
"step": 561
},
{
"epoch": 1.3353151010701545,
"grad_norm": 1.046337850818545,
"learning_rate": 1.540347782904897e-05,
"loss": 1.8589,
"step": 562
},
{
"epoch": 1.33769322235434,
"grad_norm": 1.0106090373600465,
"learning_rate": 1.5387349221574493e-05,
"loss": 2.0144,
"step": 563
},
{
"epoch": 1.3400713436385256,
"grad_norm": 1.1400911418818147,
"learning_rate": 1.5371200844993332e-05,
"loss": 1.9842,
"step": 564
},
{
"epoch": 1.342449464922711,
"grad_norm": 0.8684778766824857,
"learning_rate": 1.535503275856264e-05,
"loss": 2.0641,
"step": 565
},
{
"epoch": 1.3448275862068966,
"grad_norm": 0.9355737090349231,
"learning_rate": 1.5338845021611906e-05,
"loss": 1.8084,
"step": 566
},
{
"epoch": 1.347205707491082,
"grad_norm": 1.0306221444200083,
"learning_rate": 1.5322637693542716e-05,
"loss": 2.0131,
"step": 567
},
{
"epoch": 1.3495838287752675,
"grad_norm": 1.0214723689545668,
"learning_rate": 1.5306410833828534e-05,
"loss": 1.969,
"step": 568
},
{
"epoch": 1.351961950059453,
"grad_norm": 0.9625636615112916,
"learning_rate": 1.529016450201453e-05,
"loss": 1.9543,
"step": 569
},
{
"epoch": 1.3543400713436387,
"grad_norm": 0.9724729717579162,
"learning_rate": 1.5273898757717295e-05,
"loss": 1.929,
"step": 570
},
{
"epoch": 1.356718192627824,
"grad_norm": 1.0109447015124593,
"learning_rate": 1.5257613660624673e-05,
"loss": 1.9279,
"step": 571
},
{
"epoch": 1.3590963139120096,
"grad_norm": 1.008542064536025,
"learning_rate": 1.5241309270495524e-05,
"loss": 2.0551,
"step": 572
},
{
"epoch": 1.361474435196195,
"grad_norm": 1.2854483484165498,
"learning_rate": 1.5224985647159489e-05,
"loss": 2.0914,
"step": 573
},
{
"epoch": 1.3638525564803805,
"grad_norm": 0.9055884973780003,
"learning_rate": 1.5208642850516806e-05,
"loss": 1.8825,
"step": 574
},
{
"epoch": 1.366230677764566,
"grad_norm": 1.0375062326116673,
"learning_rate": 1.5192280940538058e-05,
"loss": 1.8674,
"step": 575
},
{
"epoch": 1.3686087990487514,
"grad_norm": 0.896010533212188,
"learning_rate": 1.5175899977263963e-05,
"loss": 2.0912,
"step": 576
},
{
"epoch": 1.370986920332937,
"grad_norm": 1.1501881622163808,
"learning_rate": 1.5159500020805173e-05,
"loss": 1.89,
"step": 577
},
{
"epoch": 1.3733650416171224,
"grad_norm": 0.9783540504581939,
"learning_rate": 1.5143081131342015e-05,
"loss": 1.9331,
"step": 578
},
{
"epoch": 1.375743162901308,
"grad_norm": 1.170097731492715,
"learning_rate": 1.5126643369124303e-05,
"loss": 2.0069,
"step": 579
},
{
"epoch": 1.3781212841854935,
"grad_norm": 1.0072974136421486,
"learning_rate": 1.5110186794471105e-05,
"loss": 2.0293,
"step": 580
},
{
"epoch": 1.380499405469679,
"grad_norm": 1.1697150901196478,
"learning_rate": 1.5093711467770517e-05,
"loss": 1.9685,
"step": 581
},
{
"epoch": 1.3828775267538644,
"grad_norm": 1.2206930372756868,
"learning_rate": 1.5077217449479455e-05,
"loss": 1.933,
"step": 582
},
{
"epoch": 1.38525564803805,
"grad_norm": 0.970392976217639,
"learning_rate": 1.5060704800123413e-05,
"loss": 2.0112,
"step": 583
},
{
"epoch": 1.3876337693222354,
"grad_norm": 1.0661843012528456,
"learning_rate": 1.5044173580296267e-05,
"loss": 1.9158,
"step": 584
},
{
"epoch": 1.390011890606421,
"grad_norm": 0.8514185489462422,
"learning_rate": 1.502762385066002e-05,
"loss": 1.8758,
"step": 585
},
{
"epoch": 1.3923900118906065,
"grad_norm": 0.9748541634753175,
"learning_rate": 1.5011055671944616e-05,
"loss": 2.1468,
"step": 586
},
{
"epoch": 1.3947681331747919,
"grad_norm": 0.8818979397080599,
"learning_rate": 1.4994469104947686e-05,
"loss": 1.9446,
"step": 587
},
{
"epoch": 1.3971462544589774,
"grad_norm": 1.167689162690369,
"learning_rate": 1.4977864210534341e-05,
"loss": 1.706,
"step": 588
},
{
"epoch": 1.3995243757431628,
"grad_norm": 0.9558567418999584,
"learning_rate": 1.496124104963695e-05,
"loss": 2.1427,
"step": 589
},
{
"epoch": 1.4019024970273484,
"grad_norm": 0.991232122819226,
"learning_rate": 1.4944599683254903e-05,
"loss": 2.0682,
"step": 590
},
{
"epoch": 1.404280618311534,
"grad_norm": 1.0316875876904197,
"learning_rate": 1.4927940172454405e-05,
"loss": 2.1085,
"step": 591
},
{
"epoch": 1.4066587395957193,
"grad_norm": 1.0732162496383806,
"learning_rate": 1.4911262578368233e-05,
"loss": 1.8279,
"step": 592
},
{
"epoch": 1.4090368608799049,
"grad_norm": 0.9257277902304547,
"learning_rate": 1.4894566962195532e-05,
"loss": 2.0153,
"step": 593
},
{
"epoch": 1.4114149821640904,
"grad_norm": 1.0669817139188487,
"learning_rate": 1.4877853385201569e-05,
"loss": 1.9106,
"step": 594
},
{
"epoch": 1.4137931034482758,
"grad_norm": 0.8934624324805909,
"learning_rate": 1.4861121908717529e-05,
"loss": 1.9279,
"step": 595
},
{
"epoch": 1.4161712247324614,
"grad_norm": 1.0784062752165995,
"learning_rate": 1.4844372594140271e-05,
"loss": 1.8346,
"step": 596
},
{
"epoch": 1.418549346016647,
"grad_norm": 1.0964764861814489,
"learning_rate": 1.4827605502932118e-05,
"loss": 2.1296,
"step": 597
},
{
"epoch": 1.4209274673008323,
"grad_norm": 15.356619023384987,
"learning_rate": 1.4810820696620625e-05,
"loss": 2.1312,
"step": 598
},
{
"epoch": 1.4233055885850179,
"grad_norm": 1.1118633779533458,
"learning_rate": 1.4794018236798349e-05,
"loss": 1.8619,
"step": 599
},
{
"epoch": 1.4256837098692032,
"grad_norm": 1.153501655154434,
"learning_rate": 1.477719818512263e-05,
"loss": 2.077,
"step": 600
},
{
"epoch": 1.4280618311533888,
"grad_norm": 1.0076021147835075,
"learning_rate": 1.4760360603315362e-05,
"loss": 1.9109,
"step": 601
},
{
"epoch": 1.4304399524375744,
"grad_norm": 0.9674703722137163,
"learning_rate": 1.4743505553162765e-05,
"loss": 2.0096,
"step": 602
},
{
"epoch": 1.4328180737217597,
"grad_norm": 1.0412828169607384,
"learning_rate": 1.4726633096515163e-05,
"loss": 1.9162,
"step": 603
},
{
"epoch": 1.4351961950059453,
"grad_norm": 0.8719798048430073,
"learning_rate": 1.4709743295286751e-05,
"loss": 1.9538,
"step": 604
},
{
"epoch": 1.4375743162901307,
"grad_norm": 1.1442918189307298,
"learning_rate": 1.4692836211455373e-05,
"loss": 1.9587,
"step": 605
},
{
"epoch": 1.4399524375743162,
"grad_norm": 1.023556986586939,
"learning_rate": 1.4675911907062289e-05,
"loss": 1.8141,
"step": 606
},
{
"epoch": 1.4423305588585018,
"grad_norm": 0.8486875820991789,
"learning_rate": 1.4658970444211953e-05,
"loss": 1.9474,
"step": 607
},
{
"epoch": 1.4447086801426874,
"grad_norm": 0.906460047909353,
"learning_rate": 1.464201188507178e-05,
"loss": 2.2903,
"step": 608
},
{
"epoch": 1.4470868014268727,
"grad_norm": 1.1791852257245115,
"learning_rate": 1.4625036291871926e-05,
"loss": 2.0229,
"step": 609
},
{
"epoch": 1.4494649227110583,
"grad_norm": 1.051198365641329,
"learning_rate": 1.460804372690505e-05,
"loss": 1.8857,
"step": 610
},
{
"epoch": 1.4518430439952437,
"grad_norm": 0.8672989481026866,
"learning_rate": 1.4591034252526093e-05,
"loss": 1.9414,
"step": 611
},
{
"epoch": 1.4542211652794292,
"grad_norm": 0.9235870505947725,
"learning_rate": 1.4574007931152037e-05,
"loss": 2.0249,
"step": 612
},
{
"epoch": 1.4565992865636148,
"grad_norm": 1.0522141111911232,
"learning_rate": 1.4556964825261696e-05,
"loss": 2.2079,
"step": 613
},
{
"epoch": 1.4589774078478002,
"grad_norm": 1.006897151013655,
"learning_rate": 1.4539904997395468e-05,
"loss": 1.9247,
"step": 614
},
{
"epoch": 1.4613555291319857,
"grad_norm": 0.9494032663031484,
"learning_rate": 1.4522828510155123e-05,
"loss": 1.9099,
"step": 615
},
{
"epoch": 1.463733650416171,
"grad_norm": 1.0076075073121573,
"learning_rate": 1.4505735426203545e-05,
"loss": 1.8666,
"step": 616
},
{
"epoch": 1.4661117717003567,
"grad_norm": 1.1408823580186684,
"learning_rate": 1.4488625808264536e-05,
"loss": 1.9044,
"step": 617
},
{
"epoch": 1.4684898929845422,
"grad_norm": 0.9085119845888443,
"learning_rate": 1.4471499719122565e-05,
"loss": 1.8932,
"step": 618
},
{
"epoch": 1.4708680142687278,
"grad_norm": 1.148726123633719,
"learning_rate": 1.4454357221622546e-05,
"loss": 1.9092,
"step": 619
},
{
"epoch": 1.4732461355529132,
"grad_norm": 0.9671542619932931,
"learning_rate": 1.4437198378669598e-05,
"loss": 1.9493,
"step": 620
},
{
"epoch": 1.4756242568370987,
"grad_norm": 1.2173093104656718,
"learning_rate": 1.4420023253228825e-05,
"loss": 2.2601,
"step": 621
},
{
"epoch": 1.478002378121284,
"grad_norm": 1.1927753831747965,
"learning_rate": 1.4402831908325082e-05,
"loss": 1.8732,
"step": 622
},
{
"epoch": 1.4803804994054697,
"grad_norm": 1.0640642993825382,
"learning_rate": 1.4385624407042729e-05,
"loss": 2.0229,
"step": 623
},
{
"epoch": 1.4827586206896552,
"grad_norm": 1.0742996295504268,
"learning_rate": 1.4368400812525434e-05,
"loss": 1.8983,
"step": 624
},
{
"epoch": 1.4851367419738406,
"grad_norm": 0.9228290280482974,
"learning_rate": 1.4351161187975904e-05,
"loss": 2.1249,
"step": 625
},
{
"epoch": 1.4875148632580262,
"grad_norm": 0.9738223203214996,
"learning_rate": 1.4333905596655668e-05,
"loss": 1.9615,
"step": 626
},
{
"epoch": 1.4898929845422115,
"grad_norm": 1.0847722352198572,
"learning_rate": 1.4316634101884859e-05,
"loss": 2.0659,
"step": 627
},
{
"epoch": 1.492271105826397,
"grad_norm": 1.2154947037501092,
"learning_rate": 1.4299346767041956e-05,
"loss": 2.0163,
"step": 628
},
{
"epoch": 1.4946492271105827,
"grad_norm": 1.1111031435058485,
"learning_rate": 1.4282043655563566e-05,
"loss": 2.1607,
"step": 629
},
{
"epoch": 1.4970273483947683,
"grad_norm": 1.225646636267743,
"learning_rate": 1.4264724830944198e-05,
"loss": 2.0063,
"step": 630
},
{
"epoch": 1.4970273483947683,
"eval_loss": 2.4099409580230713,
"eval_runtime": 65.5151,
"eval_samples_per_second": 9.753,
"eval_steps_per_second": 1.221,
"step": 630
},
{
"epoch": 1.4994054696789536,
"grad_norm": 0.9404929651596697,
"learning_rate": 1.4247390356736011e-05,
"loss": 2.1291,
"step": 631
},
{
"epoch": 1.501783590963139,
"grad_norm": 1.166127163010488,
"learning_rate": 1.4230040296548588e-05,
"loss": 1.9927,
"step": 632
},
{
"epoch": 1.5041617122473245,
"grad_norm": 0.9098808290926231,
"learning_rate": 1.4212674714048721e-05,
"loss": 2.0014,
"step": 633
},
{
"epoch": 1.50653983353151,
"grad_norm": 1.2163691555238052,
"learning_rate": 1.4195293672960148e-05,
"loss": 1.8449,
"step": 634
},
{
"epoch": 1.5089179548156957,
"grad_norm": 1.0668500527745948,
"learning_rate": 1.4177897237063336e-05,
"loss": 1.8293,
"step": 635
},
{
"epoch": 1.5112960760998813,
"grad_norm": 0.9965754757162669,
"learning_rate": 1.4160485470195245e-05,
"loss": 1.9292,
"step": 636
},
{
"epoch": 1.5136741973840666,
"grad_norm": 0.9613885332924135,
"learning_rate": 1.4143058436249095e-05,
"loss": 1.9555,
"step": 637
},
{
"epoch": 1.516052318668252,
"grad_norm": 1.029262128237135,
"learning_rate": 1.4125616199174125e-05,
"loss": 2.0374,
"step": 638
},
{
"epoch": 1.5184304399524375,
"grad_norm": 1.2261767783176603,
"learning_rate": 1.4108158822975368e-05,
"loss": 1.8905,
"step": 639
},
{
"epoch": 1.5208085612366231,
"grad_norm": 1.1691866000837525,
"learning_rate": 1.4090686371713403e-05,
"loss": 2.0864,
"step": 640
},
{
"epoch": 1.5231866825208087,
"grad_norm": 1.269895781196831,
"learning_rate": 1.4073198909504128e-05,
"loss": 1.9734,
"step": 641
},
{
"epoch": 1.525564803804994,
"grad_norm": 1.2954844306148612,
"learning_rate": 1.4055696500518539e-05,
"loss": 1.8801,
"step": 642
},
{
"epoch": 1.5279429250891794,
"grad_norm": 1.056075124876724,
"learning_rate": 1.4038179208982459e-05,
"loss": 1.9915,
"step": 643
},
{
"epoch": 1.530321046373365,
"grad_norm": 1.2282288043296659,
"learning_rate": 1.402064709917634e-05,
"loss": 1.9476,
"step": 644
},
{
"epoch": 1.5326991676575505,
"grad_norm": 1.3379065950112186,
"learning_rate": 1.4003100235434998e-05,
"loss": 2.0888,
"step": 645
},
{
"epoch": 1.5350772889417361,
"grad_norm": 0.8299776495667625,
"learning_rate": 1.3985538682147395e-05,
"loss": 1.9002,
"step": 646
},
{
"epoch": 1.5374554102259215,
"grad_norm": 1.3323866450270925,
"learning_rate": 1.3967962503756401e-05,
"loss": 1.8784,
"step": 647
},
{
"epoch": 1.539833531510107,
"grad_norm": 1.2995832056920724,
"learning_rate": 1.3950371764758543e-05,
"loss": 1.687,
"step": 648
},
{
"epoch": 1.5422116527942924,
"grad_norm": 1.0444385989486211,
"learning_rate": 1.3932766529703787e-05,
"loss": 2.0069,
"step": 649
},
{
"epoch": 1.544589774078478,
"grad_norm": 1.0757353584625595,
"learning_rate": 1.3915146863195292e-05,
"loss": 1.9444,
"step": 650
},
{
"epoch": 1.5469678953626635,
"grad_norm": 1.011252037188437,
"learning_rate": 1.3897512829889168e-05,
"loss": 1.8028,
"step": 651
},
{
"epoch": 1.5493460166468491,
"grad_norm": 1.0760288809230223,
"learning_rate": 1.3879864494494252e-05,
"loss": 2.0473,
"step": 652
},
{
"epoch": 1.5517241379310345,
"grad_norm": 1.242394869159288,
"learning_rate": 1.3862201921771864e-05,
"loss": 2.1959,
"step": 653
},
{
"epoch": 1.5541022592152198,
"grad_norm": 1.052411681271537,
"learning_rate": 1.3844525176535557e-05,
"loss": 2.0419,
"step": 654
},
{
"epoch": 1.5564803804994054,
"grad_norm": 0.9881132820543386,
"learning_rate": 1.3826834323650899e-05,
"loss": 1.8639,
"step": 655
},
{
"epoch": 1.558858501783591,
"grad_norm": 1.1464153070903584,
"learning_rate": 1.3809129428035229e-05,
"loss": 1.7864,
"step": 656
},
{
"epoch": 1.5612366230677766,
"grad_norm": 0.9309263598280921,
"learning_rate": 1.3791410554657412e-05,
"loss": 1.9641,
"step": 657
},
{
"epoch": 1.563614744351962,
"grad_norm": 1.0278232179656681,
"learning_rate": 1.3773677768537608e-05,
"loss": 2.0758,
"step": 658
},
{
"epoch": 1.5659928656361475,
"grad_norm": 1.076577593564504,
"learning_rate": 1.3755931134747024e-05,
"loss": 2.0546,
"step": 659
},
{
"epoch": 1.5683709869203328,
"grad_norm": 1.10395359678112,
"learning_rate": 1.3738170718407689e-05,
"loss": 1.9597,
"step": 660
},
{
"epoch": 1.5707491082045184,
"grad_norm": 1.0709571819780048,
"learning_rate": 1.3720396584692204e-05,
"loss": 1.8091,
"step": 661
},
{
"epoch": 1.573127229488704,
"grad_norm": 0.9361930895015704,
"learning_rate": 1.3702608798823506e-05,
"loss": 2.151,
"step": 662
},
{
"epoch": 1.5755053507728896,
"grad_norm": 0.960233741717821,
"learning_rate": 1.3684807426074637e-05,
"loss": 2.0959,
"step": 663
},
{
"epoch": 1.577883472057075,
"grad_norm": 1.0144097356398964,
"learning_rate": 1.3666992531768482e-05,
"loss": 2.0612,
"step": 664
},
{
"epoch": 1.5802615933412603,
"grad_norm": 1.0093128761118164,
"learning_rate": 1.3649164181277554e-05,
"loss": 1.9669,
"step": 665
},
{
"epoch": 1.5826397146254458,
"grad_norm": 0.8991548607233194,
"learning_rate": 1.3631322440023743e-05,
"loss": 1.9549,
"step": 666
},
{
"epoch": 1.5850178359096314,
"grad_norm": 1.1136062306803636,
"learning_rate": 1.361346737347808e-05,
"loss": 1.7156,
"step": 667
},
{
"epoch": 1.587395957193817,
"grad_norm": 1.0226580218095702,
"learning_rate": 1.359559904716048e-05,
"loss": 2.0015,
"step": 668
},
{
"epoch": 1.5897740784780023,
"grad_norm": 1.0629689645649965,
"learning_rate": 1.3577717526639532e-05,
"loss": 1.9631,
"step": 669
},
{
"epoch": 1.592152199762188,
"grad_norm": 0.9835886516605551,
"learning_rate": 1.3559822877532234e-05,
"loss": 2.0375,
"step": 670
},
{
"epoch": 1.5945303210463733,
"grad_norm": 1.0706093353458817,
"learning_rate": 1.354191516550376e-05,
"loss": 2.1751,
"step": 671
},
{
"epoch": 1.5969084423305588,
"grad_norm": 0.9758001603671903,
"learning_rate": 1.352399445626722e-05,
"loss": 2.0163,
"step": 672
},
{
"epoch": 1.5992865636147444,
"grad_norm": 0.9532313551731908,
"learning_rate": 1.3506060815583415e-05,
"loss": 1.9762,
"step": 673
},
{
"epoch": 1.60166468489893,
"grad_norm": 1.1096343614759168,
"learning_rate": 1.3488114309260603e-05,
"loss": 1.9759,
"step": 674
},
{
"epoch": 1.6040428061831153,
"grad_norm": 0.9157371697281291,
"learning_rate": 1.3470155003154251e-05,
"loss": 2.0912,
"step": 675
},
{
"epoch": 1.6064209274673007,
"grad_norm": 1.0482149472996887,
"learning_rate": 1.3452182963166792e-05,
"loss": 2.1044,
"step": 676
},
{
"epoch": 1.6087990487514863,
"grad_norm": 0.8905172826303334,
"learning_rate": 1.3434198255247399e-05,
"loss": 1.6975,
"step": 677
},
{
"epoch": 1.6111771700356718,
"grad_norm": 1.058790898557101,
"learning_rate": 1.3416200945391711e-05,
"loss": 1.8314,
"step": 678
},
{
"epoch": 1.6135552913198574,
"grad_norm": 1.0907756047211878,
"learning_rate": 1.3398191099641622e-05,
"loss": 1.7873,
"step": 679
},
{
"epoch": 1.6159334126040428,
"grad_norm": 0.9443953519393393,
"learning_rate": 1.3380168784085028e-05,
"loss": 2.0068,
"step": 680
},
{
"epoch": 1.6183115338882283,
"grad_norm": 0.8651637816722115,
"learning_rate": 1.3362134064855583e-05,
"loss": 1.8158,
"step": 681
},
{
"epoch": 1.6206896551724137,
"grad_norm": 1.0280612331110308,
"learning_rate": 1.334408700813245e-05,
"loss": 2.0307,
"step": 682
},
{
"epoch": 1.6230677764565993,
"grad_norm": 1.1338005074791748,
"learning_rate": 1.3326027680140075e-05,
"loss": 1.884,
"step": 683
},
{
"epoch": 1.6254458977407849,
"grad_norm": 0.8710520136707882,
"learning_rate": 1.3307956147147924e-05,
"loss": 1.9489,
"step": 684
},
{
"epoch": 1.6278240190249704,
"grad_norm": 1.2459709942333668,
"learning_rate": 1.3289872475470257e-05,
"loss": 2.1625,
"step": 685
},
{
"epoch": 1.6302021403091558,
"grad_norm": 0.8883559924456562,
"learning_rate": 1.3271776731465878e-05,
"loss": 1.8351,
"step": 686
},
{
"epoch": 1.6325802615933411,
"grad_norm": 0.9532577115031986,
"learning_rate": 1.325366898153788e-05,
"loss": 1.8954,
"step": 687
},
{
"epoch": 1.6349583828775267,
"grad_norm": 1.0867243605338575,
"learning_rate": 1.3235549292133425e-05,
"loss": 2.023,
"step": 688
},
{
"epoch": 1.6373365041617123,
"grad_norm": 0.9945570973943012,
"learning_rate": 1.3217417729743483e-05,
"loss": 1.7962,
"step": 689
},
{
"epoch": 1.6397146254458979,
"grad_norm": 0.9349059265184032,
"learning_rate": 1.319927436090259e-05,
"loss": 2.0278,
"step": 690
},
{
"epoch": 1.6420927467300832,
"grad_norm": 0.9423473629997007,
"learning_rate": 1.3181119252188612e-05,
"loss": 1.9311,
"step": 691
},
{
"epoch": 1.6444708680142688,
"grad_norm": 1.0039996608993265,
"learning_rate": 1.3162952470222488e-05,
"loss": 1.9779,
"step": 692
},
{
"epoch": 1.6468489892984541,
"grad_norm": 1.0292101999229477,
"learning_rate": 1.3144774081667993e-05,
"loss": 2.0786,
"step": 693
},
{
"epoch": 1.6492271105826397,
"grad_norm": 0.8987803262203654,
"learning_rate": 1.31265841532315e-05,
"loss": 1.7987,
"step": 694
},
{
"epoch": 1.6516052318668253,
"grad_norm": 1.1875414858923983,
"learning_rate": 1.3108382751661722e-05,
"loss": 2.0059,
"step": 695
},
{
"epoch": 1.6539833531510109,
"grad_norm": 1.0269203634847877,
"learning_rate": 1.3090169943749475e-05,
"loss": 2.0657,
"step": 696
},
{
"epoch": 1.6563614744351962,
"grad_norm": 1.0926982865846275,
"learning_rate": 1.3071945796327431e-05,
"loss": 2.0779,
"step": 697
},
{
"epoch": 1.6587395957193816,
"grad_norm": 1.0775345370788383,
"learning_rate": 1.3053710376269873e-05,
"loss": 1.9898,
"step": 698
},
{
"epoch": 1.6611177170035671,
"grad_norm": 1.1340932134975976,
"learning_rate": 1.3035463750492448e-05,
"loss": 2.146,
"step": 699
},
{
"epoch": 1.6634958382877527,
"grad_norm": 1.0731177791654507,
"learning_rate": 1.3017205985951926e-05,
"loss": 2.0155,
"step": 700
},
{
"epoch": 1.6658739595719383,
"grad_norm": 0.8969768193578244,
"learning_rate": 1.2998937149645944e-05,
"loss": 1.7347,
"step": 701
},
{
"epoch": 1.6682520808561236,
"grad_norm": 0.9835389016614261,
"learning_rate": 1.2980657308612778e-05,
"loss": 1.8539,
"step": 702
},
{
"epoch": 1.670630202140309,
"grad_norm": 1.1380423436787481,
"learning_rate": 1.2962366529931076e-05,
"loss": 1.9974,
"step": 703
},
{
"epoch": 1.6730083234244946,
"grad_norm": 1.0964744709498448,
"learning_rate": 1.2944064880719634e-05,
"loss": 1.72,
"step": 704
},
{
"epoch": 1.6753864447086801,
"grad_norm": 1.2519840244518319,
"learning_rate": 1.2925752428137126e-05,
"loss": 1.8856,
"step": 705
},
{
"epoch": 1.6777645659928657,
"grad_norm": 1.0121056685710188,
"learning_rate": 1.2907429239381872e-05,
"loss": 2.0765,
"step": 706
},
{
"epoch": 1.6801426872770513,
"grad_norm": 1.0586498134607165,
"learning_rate": 1.2889095381691594e-05,
"loss": 2.1218,
"step": 707
},
{
"epoch": 1.6825208085612366,
"grad_norm": 0.969809896629875,
"learning_rate": 1.287075092234316e-05,
"loss": 2.0508,
"step": 708
},
{
"epoch": 1.684898929845422,
"grad_norm": 1.0575764929431117,
"learning_rate": 1.2852395928652343e-05,
"loss": 1.9796,
"step": 709
},
{
"epoch": 1.6872770511296076,
"grad_norm": 0.9921956479219585,
"learning_rate": 1.2834030467973572e-05,
"loss": 2.0071,
"step": 710
},
{
"epoch": 1.6896551724137931,
"grad_norm": 1.0183541379675678,
"learning_rate": 1.2815654607699687e-05,
"loss": 1.9518,
"step": 711
},
{
"epoch": 1.6920332936979787,
"grad_norm": 1.097875724988254,
"learning_rate": 1.2797268415261681e-05,
"loss": 2.0982,
"step": 712
},
{
"epoch": 1.694411414982164,
"grad_norm": 0.9022331837037142,
"learning_rate": 1.2778871958128472e-05,
"loss": 1.7788,
"step": 713
},
{
"epoch": 1.6967895362663494,
"grad_norm": 0.8133273901988929,
"learning_rate": 1.2760465303806639e-05,
"loss": 1.8453,
"step": 714
},
{
"epoch": 1.699167657550535,
"grad_norm": 1.0492660319739247,
"learning_rate": 1.2742048519840182e-05,
"loss": 1.8965,
"step": 715
},
{
"epoch": 1.7015457788347206,
"grad_norm": 1.1192820098120624,
"learning_rate": 1.2723621673810277e-05,
"loss": 2.0887,
"step": 716
},
{
"epoch": 1.7039239001189062,
"grad_norm": 0.9360576963110526,
"learning_rate": 1.2705184833335008e-05,
"loss": 1.7144,
"step": 717
},
{
"epoch": 1.7063020214030915,
"grad_norm": 1.1130603696459762,
"learning_rate": 1.2686738066069148e-05,
"loss": 1.7416,
"step": 718
},
{
"epoch": 1.708680142687277,
"grad_norm": 1.0304078304302904,
"learning_rate": 1.2668281439703893e-05,
"loss": 2.1117,
"step": 719
},
{
"epoch": 1.7110582639714624,
"grad_norm": 1.0656197340261881,
"learning_rate": 1.264981502196662e-05,
"loss": 1.9695,
"step": 720
},
{
"epoch": 1.713436385255648,
"grad_norm": 1.0087951542362363,
"learning_rate": 1.2631338880620627e-05,
"loss": 1.9352,
"step": 721
},
{
"epoch": 1.7158145065398336,
"grad_norm": 1.0373066800416066,
"learning_rate": 1.2612853083464903e-05,
"loss": 2.0947,
"step": 722
},
{
"epoch": 1.7181926278240192,
"grad_norm": 1.1765591294813367,
"learning_rate": 1.2594357698333864e-05,
"loss": 2.0922,
"step": 723
},
{
"epoch": 1.7205707491082045,
"grad_norm": 1.1664325067262717,
"learning_rate": 1.2575852793097113e-05,
"loss": 2.1225,
"step": 724
},
{
"epoch": 1.7229488703923899,
"grad_norm": 1.229465533643844,
"learning_rate": 1.255733843565918e-05,
"loss": 1.861,
"step": 725
},
{
"epoch": 1.7253269916765754,
"grad_norm": 0.9938570044256105,
"learning_rate": 1.2538814693959294e-05,
"loss": 1.9642,
"step": 726
},
{
"epoch": 1.727705112960761,
"grad_norm": 1.0606410107508442,
"learning_rate": 1.2520281635971103e-05,
"loss": 1.9007,
"step": 727
},
{
"epoch": 1.7300832342449466,
"grad_norm": 1.064544615609171,
"learning_rate": 1.2501739329702453e-05,
"loss": 2.0957,
"step": 728
},
{
"epoch": 1.732461355529132,
"grad_norm": 1.2317528483018596,
"learning_rate": 1.2483187843195128e-05,
"loss": 2.0826,
"step": 729
},
{
"epoch": 1.7348394768133175,
"grad_norm": 0.8079713987023232,
"learning_rate": 1.2464627244524595e-05,
"loss": 1.9538,
"step": 730
},
{
"epoch": 1.7372175980975029,
"grad_norm": 1.2758113698136755,
"learning_rate": 1.2446057601799753e-05,
"loss": 1.9805,
"step": 731
},
{
"epoch": 1.7395957193816884,
"grad_norm": 1.0629931764775897,
"learning_rate": 1.2427478983162694e-05,
"loss": 1.9351,
"step": 732
},
{
"epoch": 1.741973840665874,
"grad_norm": 0.9152635150811873,
"learning_rate": 1.2408891456788457e-05,
"loss": 1.8034,
"step": 733
},
{
"epoch": 1.7443519619500596,
"grad_norm": 1.1182721783403513,
"learning_rate": 1.2390295090884752e-05,
"loss": 2.0279,
"step": 734
},
{
"epoch": 1.746730083234245,
"grad_norm": 0.8598255371847341,
"learning_rate": 1.2371689953691733e-05,
"loss": 1.8253,
"step": 735
},
{
"epoch": 1.746730083234245,
"eval_loss": 2.387751340866089,
"eval_runtime": 66.4359,
"eval_samples_per_second": 9.618,
"eval_steps_per_second": 1.204,
"step": 735
},
{
"epoch": 1.7491082045184303,
"grad_norm": 0.9861227291838077,
"learning_rate": 1.2353076113481742e-05,
"loss": 1.9038,
"step": 736
},
{
"epoch": 1.7514863258026159,
"grad_norm": 1.039133223415763,
"learning_rate": 1.2334453638559057e-05,
"loss": 2.0651,
"step": 737
},
{
"epoch": 1.7538644470868014,
"grad_norm": 1.036263407591309,
"learning_rate": 1.2315822597259636e-05,
"loss": 1.7324,
"step": 738
},
{
"epoch": 1.756242568370987,
"grad_norm": 0.9557591103456263,
"learning_rate": 1.2297183057950886e-05,
"loss": 2.0171,
"step": 739
},
{
"epoch": 1.7586206896551724,
"grad_norm": 0.9906419044217084,
"learning_rate": 1.2278535089031377e-05,
"loss": 1.8361,
"step": 740
},
{
"epoch": 1.760998810939358,
"grad_norm": 1.0631674141525231,
"learning_rate": 1.2259878758930627e-05,
"loss": 1.9345,
"step": 741
},
{
"epoch": 1.7633769322235433,
"grad_norm": 1.005919565429525,
"learning_rate": 1.2241214136108834e-05,
"loss": 2.0726,
"step": 742
},
{
"epoch": 1.7657550535077289,
"grad_norm": 0.8171183104600345,
"learning_rate": 1.222254128905662e-05,
"loss": 1.9147,
"step": 743
},
{
"epoch": 1.7681331747919145,
"grad_norm": 1.025525655835034,
"learning_rate": 1.220386028629479e-05,
"loss": 1.9566,
"step": 744
},
{
"epoch": 1.7705112960761,
"grad_norm": 1.1199177387750694,
"learning_rate": 1.218517119637408e-05,
"loss": 1.8823,
"step": 745
},
{
"epoch": 1.7728894173602854,
"grad_norm": 0.8074812474679398,
"learning_rate": 1.2166474087874893e-05,
"loss": 1.8039,
"step": 746
},
{
"epoch": 1.7752675386444707,
"grad_norm": 0.8574176753575323,
"learning_rate": 1.2147769029407069e-05,
"loss": 1.911,
"step": 747
},
{
"epoch": 1.7776456599286563,
"grad_norm": 1.0390138098655621,
"learning_rate": 1.212905608960961e-05,
"loss": 1.6973,
"step": 748
},
{
"epoch": 1.7800237812128419,
"grad_norm": 0.9501509089266569,
"learning_rate": 1.2110335337150445e-05,
"loss": 2.0167,
"step": 749
},
{
"epoch": 1.7824019024970275,
"grad_norm": 1.021610077627784,
"learning_rate": 1.209160684072617e-05,
"loss": 2.032,
"step": 750
},
{
"epoch": 1.7847800237812128,
"grad_norm": 1.067584580316804,
"learning_rate": 1.2072870669061793e-05,
"loss": 2.0627,
"step": 751
},
{
"epoch": 1.7871581450653984,
"grad_norm": 0.971924208763616,
"learning_rate": 1.2054126890910499e-05,
"loss": 1.9211,
"step": 752
},
{
"epoch": 1.7895362663495837,
"grad_norm": 1.059861690139045,
"learning_rate": 1.2035375575053373e-05,
"loss": 1.8785,
"step": 753
},
{
"epoch": 1.7919143876337693,
"grad_norm": 0.9537414827635367,
"learning_rate": 1.2016616790299165e-05,
"loss": 1.9083,
"step": 754
},
{
"epoch": 1.7942925089179549,
"grad_norm": 0.9591859502901945,
"learning_rate": 1.1997850605484034e-05,
"loss": 2.0818,
"step": 755
},
{
"epoch": 1.7966706302021405,
"grad_norm": 1.0432501587679868,
"learning_rate": 1.1979077089471288e-05,
"loss": 1.9283,
"step": 756
},
{
"epoch": 1.7990487514863258,
"grad_norm": 0.9516347204385417,
"learning_rate": 1.1960296311151143e-05,
"loss": 2.0293,
"step": 757
},
{
"epoch": 1.8014268727705112,
"grad_norm": 1.0424747053382684,
"learning_rate": 1.1941508339440467e-05,
"loss": 2.0464,
"step": 758
},
{
"epoch": 1.8038049940546967,
"grad_norm": 1.0574987855006235,
"learning_rate": 1.1922713243282514e-05,
"loss": 2.0236,
"step": 759
},
{
"epoch": 1.8061831153388823,
"grad_norm": 1.0377956706796805,
"learning_rate": 1.1903911091646684e-05,
"loss": 1.9934,
"step": 760
},
{
"epoch": 1.808561236623068,
"grad_norm": 1.0769312593664324,
"learning_rate": 1.1885101953528279e-05,
"loss": 2.1783,
"step": 761
},
{
"epoch": 1.8109393579072532,
"grad_norm": 0.9429126781570604,
"learning_rate": 1.1866285897948227e-05,
"loss": 1.899,
"step": 762
},
{
"epoch": 1.8133174791914386,
"grad_norm": 1.080323404458619,
"learning_rate": 1.1847462993952842e-05,
"loss": 1.9386,
"step": 763
},
{
"epoch": 1.8156956004756242,
"grad_norm": 0.9840126348452174,
"learning_rate": 1.1828633310613569e-05,
"loss": 2.0926,
"step": 764
},
{
"epoch": 1.8180737217598097,
"grad_norm": 1.1925367572998558,
"learning_rate": 1.180979691702673e-05,
"loss": 1.8951,
"step": 765
},
{
"epoch": 1.8204518430439953,
"grad_norm": 0.9952376958926185,
"learning_rate": 1.1790953882313271e-05,
"loss": 1.9651,
"step": 766
},
{
"epoch": 1.822829964328181,
"grad_norm": 0.936960021257687,
"learning_rate": 1.1772104275618512e-05,
"loss": 1.9273,
"step": 767
},
{
"epoch": 1.8252080856123662,
"grad_norm": 0.9274559160209536,
"learning_rate": 1.175324816611188e-05,
"loss": 1.9401,
"step": 768
},
{
"epoch": 1.8275862068965516,
"grad_norm": 1.0137182594497265,
"learning_rate": 1.1734385622986668e-05,
"loss": 2.0039,
"step": 769
},
{
"epoch": 1.8299643281807372,
"grad_norm": 1.162702633094251,
"learning_rate": 1.1715516715459784e-05,
"loss": 1.993,
"step": 770
},
{
"epoch": 1.8323424494649228,
"grad_norm": 0.9064383127278726,
"learning_rate": 1.1696641512771482e-05,
"loss": 2.0824,
"step": 771
},
{
"epoch": 1.8347205707491083,
"grad_norm": 0.9662577578484263,
"learning_rate": 1.1677760084185123e-05,
"loss": 1.7069,
"step": 772
},
{
"epoch": 1.8370986920332937,
"grad_norm": 0.8709375549819791,
"learning_rate": 1.1658872498986905e-05,
"loss": 2.0223,
"step": 773
},
{
"epoch": 1.839476813317479,
"grad_norm": 1.145853044079966,
"learning_rate": 1.1639978826485629e-05,
"loss": 1.9748,
"step": 774
},
{
"epoch": 1.8418549346016646,
"grad_norm": 1.1254237494407062,
"learning_rate": 1.1621079136012426e-05,
"loss": 1.9803,
"step": 775
},
{
"epoch": 1.8442330558858502,
"grad_norm": 1.10543629656252,
"learning_rate": 1.160217349692051e-05,
"loss": 1.9952,
"step": 776
},
{
"epoch": 1.8466111771700358,
"grad_norm": 1.0421494399227835,
"learning_rate": 1.1583261978584934e-05,
"loss": 2.2644,
"step": 777
},
{
"epoch": 1.8489892984542213,
"grad_norm": 1.157252899286374,
"learning_rate": 1.156434465040231e-05,
"loss": 1.8651,
"step": 778
},
{
"epoch": 1.8513674197384067,
"grad_norm": 1.1787150544248581,
"learning_rate": 1.1545421581790579e-05,
"loss": 2.006,
"step": 779
},
{
"epoch": 1.853745541022592,
"grad_norm": 0.9108745110078712,
"learning_rate": 1.1526492842188746e-05,
"loss": 1.9716,
"step": 780
},
{
"epoch": 1.8561236623067776,
"grad_norm": 1.0584595603639568,
"learning_rate": 1.1507558501056622e-05,
"loss": 2.0011,
"step": 781
},
{
"epoch": 1.8585017835909632,
"grad_norm": 1.0921406194015688,
"learning_rate": 1.1488618627874581e-05,
"loss": 1.8896,
"step": 782
},
{
"epoch": 1.8608799048751488,
"grad_norm": 1.0068894499355592,
"learning_rate": 1.1469673292143284e-05,
"loss": 1.9903,
"step": 783
},
{
"epoch": 1.8632580261593341,
"grad_norm": 1.0157407453051925,
"learning_rate": 1.145072256338345e-05,
"loss": 1.8454,
"step": 784
},
{
"epoch": 1.8656361474435195,
"grad_norm": 1.0130783012738553,
"learning_rate": 1.1431766511135581e-05,
"loss": 1.856,
"step": 785
},
{
"epoch": 1.868014268727705,
"grad_norm": 1.1047664955736678,
"learning_rate": 1.1412805204959719e-05,
"loss": 1.881,
"step": 786
},
{
"epoch": 1.8703923900118906,
"grad_norm": 0.9130130901731418,
"learning_rate": 1.1393838714435176e-05,
"loss": 1.9902,
"step": 787
},
{
"epoch": 1.8727705112960762,
"grad_norm": 0.8444730119781477,
"learning_rate": 1.1374867109160295e-05,
"loss": 1.8033,
"step": 788
},
{
"epoch": 1.8751486325802615,
"grad_norm": 1.0068775645787829,
"learning_rate": 1.135589045875219e-05,
"loss": 1.8523,
"step": 789
},
{
"epoch": 1.8775267538644471,
"grad_norm": 1.0137172293384147,
"learning_rate": 1.1336908832846485e-05,
"loss": 1.8672,
"step": 790
},
{
"epoch": 1.8799048751486325,
"grad_norm": 1.1230734224665218,
"learning_rate": 1.1317922301097064e-05,
"loss": 1.9669,
"step": 791
},
{
"epoch": 1.882282996432818,
"grad_norm": 0.8743774773477322,
"learning_rate": 1.1298930933175805e-05,
"loss": 1.7966,
"step": 792
},
{
"epoch": 1.8846611177170036,
"grad_norm": 1.321843260348692,
"learning_rate": 1.1279934798772344e-05,
"loss": 1.9052,
"step": 793
},
{
"epoch": 1.8870392390011892,
"grad_norm": 0.9265508467848211,
"learning_rate": 1.12609339675938e-05,
"loss": 2.0626,
"step": 794
},
{
"epoch": 1.8894173602853745,
"grad_norm": 1.065099635395474,
"learning_rate": 1.1241928509364533e-05,
"loss": 1.8472,
"step": 795
},
{
"epoch": 1.89179548156956,
"grad_norm": 0.9336444235822928,
"learning_rate": 1.1222918493825876e-05,
"loss": 2.0588,
"step": 796
},
{
"epoch": 1.8941736028537455,
"grad_norm": 0.9901340362445608,
"learning_rate": 1.1203903990735888e-05,
"loss": 1.969,
"step": 797
},
{
"epoch": 1.896551724137931,
"grad_norm": 1.1007028104085261,
"learning_rate": 1.1184885069869096e-05,
"loss": 1.8668,
"step": 798
},
{
"epoch": 1.8989298454221166,
"grad_norm": 0.8704786291588872,
"learning_rate": 1.1165861801016235e-05,
"loss": 1.7756,
"step": 799
},
{
"epoch": 1.901307966706302,
"grad_norm": 0.9586858624614886,
"learning_rate": 1.1146834253984008e-05,
"loss": 2.1082,
"step": 800
},
{
"epoch": 1.9036860879904876,
"grad_norm": 0.9935316309566465,
"learning_rate": 1.1127802498594792e-05,
"loss": 2.0898,
"step": 801
},
{
"epoch": 1.906064209274673,
"grad_norm": 1.0128018872894315,
"learning_rate": 1.110876660468643e-05,
"loss": 2.0126,
"step": 802
},
{
"epoch": 1.9084423305588585,
"grad_norm": 1.0425566210644994,
"learning_rate": 1.108972664211194e-05,
"loss": 1.8575,
"step": 803
},
{
"epoch": 1.910820451843044,
"grad_norm": 1.155933364218014,
"learning_rate": 1.1070682680739275e-05,
"loss": 1.9965,
"step": 804
},
{
"epoch": 1.9131985731272296,
"grad_norm": 1.4109307764012204,
"learning_rate": 1.105163479045106e-05,
"loss": 1.7907,
"step": 805
},
{
"epoch": 1.915576694411415,
"grad_norm": 1.019545311598646,
"learning_rate": 1.1032583041144334e-05,
"loss": 1.7445,
"step": 806
},
{
"epoch": 1.9179548156956003,
"grad_norm": 0.8852176465494631,
"learning_rate": 1.1013527502730301e-05,
"loss": 1.8363,
"step": 807
},
{
"epoch": 1.920332936979786,
"grad_norm": 0.967544348388286,
"learning_rate": 1.0994468245134071e-05,
"loss": 1.7991,
"step": 808
},
{
"epoch": 1.9227110582639715,
"grad_norm": 1.005400241628268,
"learning_rate": 1.0975405338294399e-05,
"loss": 1.8783,
"step": 809
},
{
"epoch": 1.925089179548157,
"grad_norm": 0.8762088510461341,
"learning_rate": 1.0956338852163424e-05,
"loss": 1.8432,
"step": 810
},
{
"epoch": 1.9274673008323424,
"grad_norm": 1.0033258791075268,
"learning_rate": 1.0937268856706435e-05,
"loss": 1.9274,
"step": 811
},
{
"epoch": 1.929845422116528,
"grad_norm": 0.9911088707682482,
"learning_rate": 1.0918195421901583e-05,
"loss": 1.7392,
"step": 812
},
{
"epoch": 1.9322235434007133,
"grad_norm": 1.2905322005763973,
"learning_rate": 1.089911861773965e-05,
"loss": 2.1347,
"step": 813
},
{
"epoch": 1.934601664684899,
"grad_norm": 1.014836431956825,
"learning_rate": 1.0880038514223778e-05,
"loss": 1.9524,
"step": 814
},
{
"epoch": 1.9369797859690845,
"grad_norm": 0.9524664386824049,
"learning_rate": 1.0860955181369219e-05,
"loss": 1.9656,
"step": 815
},
{
"epoch": 1.93935790725327,
"grad_norm": 1.1264640139375,
"learning_rate": 1.0841868689203072e-05,
"loss": 1.8089,
"step": 816
},
{
"epoch": 1.9417360285374554,
"grad_norm": 0.8635336405252807,
"learning_rate": 1.0822779107764028e-05,
"loss": 1.908,
"step": 817
},
{
"epoch": 1.9441141498216408,
"grad_norm": 1.0677562500326228,
"learning_rate": 1.0803686507102118e-05,
"loss": 1.9362,
"step": 818
},
{
"epoch": 1.9464922711058263,
"grad_norm": 1.0413287807888576,
"learning_rate": 1.0784590957278452e-05,
"loss": 1.8287,
"step": 819
},
{
"epoch": 1.948870392390012,
"grad_norm": 1.1002052999274758,
"learning_rate": 1.076549252836496e-05,
"loss": 2.0404,
"step": 820
},
{
"epoch": 1.9512485136741975,
"grad_norm": 1.0659689575025673,
"learning_rate": 1.0746391290444136e-05,
"loss": 2.0779,
"step": 821
},
{
"epoch": 1.9536266349583828,
"grad_norm": 1.0634086392515605,
"learning_rate": 1.0727287313608783e-05,
"loss": 1.9018,
"step": 822
},
{
"epoch": 1.9560047562425684,
"grad_norm": 0.8849872440417224,
"learning_rate": 1.0708180667961758e-05,
"loss": 2.1171,
"step": 823
},
{
"epoch": 1.9583828775267538,
"grad_norm": 1.0283232728798035,
"learning_rate": 1.0689071423615708e-05,
"loss": 1.8658,
"step": 824
},
{
"epoch": 1.9607609988109393,
"grad_norm": 1.1823895638000625,
"learning_rate": 1.0669959650692818e-05,
"loss": 1.6882,
"step": 825
},
{
"epoch": 1.963139120095125,
"grad_norm": 1.1175568913711953,
"learning_rate": 1.0650845419324544e-05,
"loss": 1.855,
"step": 826
},
{
"epoch": 1.9655172413793105,
"grad_norm": 1.1061515815620258,
"learning_rate": 1.0631728799651374e-05,
"loss": 2.0233,
"step": 827
},
{
"epoch": 1.9678953626634959,
"grad_norm": 0.9330753316753481,
"learning_rate": 1.0612609861822556e-05,
"loss": 2.0104,
"step": 828
},
{
"epoch": 1.9702734839476812,
"grad_norm": 1.0424417166124378,
"learning_rate": 1.059348867599584e-05,
"loss": 1.9829,
"step": 829
},
{
"epoch": 1.9726516052318668,
"grad_norm": 0.9251278544848563,
"learning_rate": 1.0574365312337235e-05,
"loss": 1.6856,
"step": 830
},
{
"epoch": 1.9750297265160524,
"grad_norm": 1.0217360426495539,
"learning_rate": 1.055523984102073e-05,
"loss": 1.9015,
"step": 831
},
{
"epoch": 1.977407847800238,
"grad_norm": 1.01682991161189,
"learning_rate": 1.0536112332228057e-05,
"loss": 1.917,
"step": 832
},
{
"epoch": 1.9797859690844233,
"grad_norm": 0.9565389300837592,
"learning_rate": 1.0516982856148421e-05,
"loss": 1.8118,
"step": 833
},
{
"epoch": 1.9821640903686086,
"grad_norm": 1.104212144557824,
"learning_rate": 1.0497851482978247e-05,
"loss": 1.8446,
"step": 834
},
{
"epoch": 1.9845422116527942,
"grad_norm": 0.9911600279737486,
"learning_rate": 1.0478718282920922e-05,
"loss": 1.9056,
"step": 835
},
{
"epoch": 1.9869203329369798,
"grad_norm": 0.9981165586294449,
"learning_rate": 1.0459583326186532e-05,
"loss": 2.019,
"step": 836
},
{
"epoch": 1.9892984542211654,
"grad_norm": 1.2857901353617724,
"learning_rate": 1.0440446682991617e-05,
"loss": 1.8977,
"step": 837
},
{
"epoch": 1.991676575505351,
"grad_norm": 1.2212775483605185,
"learning_rate": 1.0421308423558898e-05,
"loss": 1.9891,
"step": 838
},
{
"epoch": 1.9940546967895363,
"grad_norm": 1.0720928120882365,
"learning_rate": 1.0402168618117038e-05,
"loss": 2.1227,
"step": 839
},
{
"epoch": 1.9964328180737216,
"grad_norm": 1.0906773289299942,
"learning_rate": 1.0383027336900356e-05,
"loss": 1.9278,
"step": 840
},
{
"epoch": 1.9964328180737216,
"eval_loss": 2.3622491359710693,
"eval_runtime": 65.5731,
"eval_samples_per_second": 9.745,
"eval_steps_per_second": 1.22,
"step": 840
},
{
"epoch": 1.9988109393579072,
"grad_norm": 0.9597175384609758,
"learning_rate": 1.0363884650148601e-05,
"loss": 2.1366,
"step": 841
},
{
"epoch": 2.0,
"grad_norm": 1.6494060559715102,
"learning_rate": 1.0344740628106673e-05,
"loss": 2.0977,
"step": 842
},
{
"epoch": 2.0023781212841856,
"grad_norm": 1.8572000704827933,
"learning_rate": 1.0325595341024377e-05,
"loss": 1.804,
"step": 843
},
{
"epoch": 2.004756242568371,
"grad_norm": 1.6109351938216152,
"learning_rate": 1.0306448859156155e-05,
"loss": 1.6572,
"step": 844
},
{
"epoch": 2.0071343638525563,
"grad_norm": 1.682163858699801,
"learning_rate": 1.0287301252760833e-05,
"loss": 1.9179,
"step": 845
},
{
"epoch": 2.009512485136742,
"grad_norm": 1.4642932185295525,
"learning_rate": 1.0268152592101366e-05,
"loss": 1.7944,
"step": 846
},
{
"epoch": 2.0118906064209274,
"grad_norm": 2.298918914235818,
"learning_rate": 1.024900294744458e-05,
"loss": 1.4883,
"step": 847
},
{
"epoch": 2.014268727705113,
"grad_norm": 2.1569642060668723,
"learning_rate": 1.0229852389060905e-05,
"loss": 1.7754,
"step": 848
},
{
"epoch": 2.0166468489892986,
"grad_norm": 1.650877681094919,
"learning_rate": 1.0210700987224134e-05,
"loss": 1.7676,
"step": 849
},
{
"epoch": 2.019024970273484,
"grad_norm": 1.4229362966156138,
"learning_rate": 1.0191548812211143e-05,
"loss": 1.6903,
"step": 850
},
{
"epoch": 2.0214030915576693,
"grad_norm": 1.3427234695472385,
"learning_rate": 1.0172395934301653e-05,
"loss": 1.8036,
"step": 851
},
{
"epoch": 2.023781212841855,
"grad_norm": 1.4370215727780862,
"learning_rate": 1.0153242423777964e-05,
"loss": 1.6991,
"step": 852
},
{
"epoch": 2.0261593341260404,
"grad_norm": 1.2143026318203651,
"learning_rate": 1.0134088350924699e-05,
"loss": 1.7004,
"step": 853
},
{
"epoch": 2.028537455410226,
"grad_norm": 1.2634866035152414,
"learning_rate": 1.0114933786028534e-05,
"loss": 1.7919,
"step": 854
},
{
"epoch": 2.0309155766944116,
"grad_norm": 1.3125463113477043,
"learning_rate": 1.009577879937796e-05,
"loss": 1.8493,
"step": 855
},
{
"epoch": 2.0332936979785967,
"grad_norm": 1.1854528046002857,
"learning_rate": 1.0076623461263017e-05,
"loss": 1.7386,
"step": 856
},
{
"epoch": 2.0356718192627823,
"grad_norm": 1.2742920280289385,
"learning_rate": 1.005746784197503e-05,
"loss": 1.5267,
"step": 857
},
{
"epoch": 2.038049940546968,
"grad_norm": 1.4794216244383256,
"learning_rate": 1.003831201180636e-05,
"loss": 1.68,
"step": 858
},
{
"epoch": 2.0404280618311534,
"grad_norm": 1.1043401807863704,
"learning_rate": 1.0019156041050134e-05,
"loss": 1.7469,
"step": 859
},
{
"epoch": 2.042806183115339,
"grad_norm": 1.3043477136558799,
"learning_rate": 1e-05,
"loss": 1.7495,
"step": 860
},
{
"epoch": 2.0451843043995246,
"grad_norm": 1.1439412634195671,
"learning_rate": 9.98084395894987e-06,
"loss": 1.7956,
"step": 861
},
{
"epoch": 2.0475624256837097,
"grad_norm": 1.1847513743650282,
"learning_rate": 9.961687988193642e-06,
"loss": 1.6905,
"step": 862
},
{
"epoch": 2.0499405469678953,
"grad_norm": 1.230107373003021,
"learning_rate": 9.942532158024971e-06,
"loss": 1.637,
"step": 863
},
{
"epoch": 2.052318668252081,
"grad_norm": 1.331643711464931,
"learning_rate": 9.923376538736985e-06,
"loss": 1.9403,
"step": 864
},
{
"epoch": 2.0546967895362664,
"grad_norm": 1.3885388021890999,
"learning_rate": 9.904221200622042e-06,
"loss": 1.7906,
"step": 865
},
{
"epoch": 2.057074910820452,
"grad_norm": 1.204006509857973,
"learning_rate": 9.885066213971471e-06,
"loss": 1.5899,
"step": 866
},
{
"epoch": 2.059453032104637,
"grad_norm": 1.1475843090693567,
"learning_rate": 9.865911649075305e-06,
"loss": 1.7476,
"step": 867
},
{
"epoch": 2.0618311533888227,
"grad_norm": 1.5532196494457342,
"learning_rate": 9.846757576222038e-06,
"loss": 2.0227,
"step": 868
},
{
"epoch": 2.0642092746730083,
"grad_norm": 1.1040634362125452,
"learning_rate": 9.82760406569835e-06,
"loss": 1.7182,
"step": 869
},
{
"epoch": 2.066587395957194,
"grad_norm": 1.2299107164007779,
"learning_rate": 9.80845118778886e-06,
"loss": 1.8007,
"step": 870
},
{
"epoch": 2.0689655172413794,
"grad_norm": 1.3188262031162163,
"learning_rate": 9.78929901277587e-06,
"loss": 1.8093,
"step": 871
},
{
"epoch": 2.0713436385255646,
"grad_norm": 1.303921293625063,
"learning_rate": 9.770147610939098e-06,
"loss": 1.3981,
"step": 872
},
{
"epoch": 2.07372175980975,
"grad_norm": 1.1390122295999698,
"learning_rate": 9.750997052555423e-06,
"loss": 1.6957,
"step": 873
},
{
"epoch": 2.0760998810939357,
"grad_norm": 1.7116484489773238,
"learning_rate": 9.731847407898637e-06,
"loss": 1.7392,
"step": 874
},
{
"epoch": 2.0784780023781213,
"grad_norm": 1.2294423692022918,
"learning_rate": 9.712698747239172e-06,
"loss": 1.7854,
"step": 875
},
{
"epoch": 2.080856123662307,
"grad_norm": 1.2941804638097063,
"learning_rate": 9.693551140843848e-06,
"loss": 1.6763,
"step": 876
},
{
"epoch": 2.0832342449464925,
"grad_norm": 1.167524044387485,
"learning_rate": 9.674404658975627e-06,
"loss": 1.7381,
"step": 877
},
{
"epoch": 2.0856123662306776,
"grad_norm": 1.3380153711442146,
"learning_rate": 9.655259371893329e-06,
"loss": 1.7105,
"step": 878
},
{
"epoch": 2.087990487514863,
"grad_norm": 1.3080896822024415,
"learning_rate": 9.636115349851402e-06,
"loss": 1.7833,
"step": 879
},
{
"epoch": 2.0903686087990487,
"grad_norm": 1.2366799144892753,
"learning_rate": 9.616972663099648e-06,
"loss": 1.679,
"step": 880
},
{
"epoch": 2.0927467300832343,
"grad_norm": 1.1601075579319895,
"learning_rate": 9.597831381882966e-06,
"loss": 1.6583,
"step": 881
},
{
"epoch": 2.09512485136742,
"grad_norm": 1.2534594907956864,
"learning_rate": 9.578691576441104e-06,
"loss": 1.4498,
"step": 882
},
{
"epoch": 2.097502972651605,
"grad_norm": 1.3319983592609854,
"learning_rate": 9.559553317008386e-06,
"loss": 1.791,
"step": 883
},
{
"epoch": 2.0998810939357906,
"grad_norm": 1.1817544740271848,
"learning_rate": 9.540416673813471e-06,
"loss": 1.6665,
"step": 884
},
{
"epoch": 2.102259215219976,
"grad_norm": 1.3535125184572838,
"learning_rate": 9.521281717079082e-06,
"loss": 1.6581,
"step": 885
},
{
"epoch": 2.1046373365041617,
"grad_norm": 1.3813623473085552,
"learning_rate": 9.502148517021757e-06,
"loss": 1.9743,
"step": 886
},
{
"epoch": 2.1070154577883473,
"grad_norm": 1.1982300911541166,
"learning_rate": 9.483017143851582e-06,
"loss": 1.6314,
"step": 887
},
{
"epoch": 2.109393579072533,
"grad_norm": 1.270544773734369,
"learning_rate": 9.463887667771946e-06,
"loss": 1.559,
"step": 888
},
{
"epoch": 2.111771700356718,
"grad_norm": 1.1811810082167495,
"learning_rate": 9.444760158979272e-06,
"loss": 1.6682,
"step": 889
},
{
"epoch": 2.1141498216409036,
"grad_norm": 1.1438900433747767,
"learning_rate": 9.425634687662768e-06,
"loss": 1.6899,
"step": 890
},
{
"epoch": 2.116527942925089,
"grad_norm": 1.2892989413709668,
"learning_rate": 9.406511324004162e-06,
"loss": 1.7346,
"step": 891
},
{
"epoch": 2.1189060642092747,
"grad_norm": 1.1712789054414114,
"learning_rate": 9.387390138177447e-06,
"loss": 1.6321,
"step": 892
},
{
"epoch": 2.1212841854934603,
"grad_norm": 1.1012248650637824,
"learning_rate": 9.368271200348627e-06,
"loss": 1.5844,
"step": 893
},
{
"epoch": 2.1236623067776454,
"grad_norm": 1.2478720681495163,
"learning_rate": 9.349154580675457e-06,
"loss": 1.4902,
"step": 894
},
{
"epoch": 2.126040428061831,
"grad_norm": 1.268299866629765,
"learning_rate": 9.330040349307185e-06,
"loss": 1.7093,
"step": 895
},
{
"epoch": 2.1284185493460166,
"grad_norm": 1.2311038089588855,
"learning_rate": 9.310928576384293e-06,
"loss": 1.6951,
"step": 896
},
{
"epoch": 2.130796670630202,
"grad_norm": 1.3174407079124912,
"learning_rate": 9.291819332038243e-06,
"loss": 1.8443,
"step": 897
},
{
"epoch": 2.1331747919143877,
"grad_norm": 1.2906831927984828,
"learning_rate": 9.272712686391219e-06,
"loss": 1.8221,
"step": 898
},
{
"epoch": 2.1355529131985733,
"grad_norm": 1.2522791578826056,
"learning_rate": 9.253608709555869e-06,
"loss": 1.6128,
"step": 899
},
{
"epoch": 2.1379310344827585,
"grad_norm": 1.0934517484029729,
"learning_rate": 9.234507471635043e-06,
"loss": 1.6859,
"step": 900
},
{
"epoch": 2.140309155766944,
"grad_norm": 1.194959938259093,
"learning_rate": 9.215409042721553e-06,
"loss": 1.7118,
"step": 901
},
{
"epoch": 2.1426872770511296,
"grad_norm": 1.1113101873824878,
"learning_rate": 9.196313492897885e-06,
"loss": 1.7543,
"step": 902
},
{
"epoch": 2.145065398335315,
"grad_norm": 1.207090436132185,
"learning_rate": 9.177220892235976e-06,
"loss": 1.8335,
"step": 903
},
{
"epoch": 2.1474435196195008,
"grad_norm": 1.1323648129867685,
"learning_rate": 9.15813131079693e-06,
"loss": 1.622,
"step": 904
},
{
"epoch": 2.149821640903686,
"grad_norm": 1.1631703384520493,
"learning_rate": 9.139044818630784e-06,
"loss": 1.734,
"step": 905
},
{
"epoch": 2.1521997621878715,
"grad_norm": 1.1629589692582991,
"learning_rate": 9.119961485776223e-06,
"loss": 1.6643,
"step": 906
},
{
"epoch": 2.154577883472057,
"grad_norm": 1.1621837317471908,
"learning_rate": 9.100881382260353e-06,
"loss": 1.6431,
"step": 907
},
{
"epoch": 2.1569560047562426,
"grad_norm": 1.1112908601843028,
"learning_rate": 9.08180457809842e-06,
"loss": 1.7269,
"step": 908
},
{
"epoch": 2.159334126040428,
"grad_norm": 1.2755784245054267,
"learning_rate": 9.062731143293569e-06,
"loss": 1.846,
"step": 909
},
{
"epoch": 2.1617122473246138,
"grad_norm": 1.1803023595907047,
"learning_rate": 9.043661147836578e-06,
"loss": 1.6091,
"step": 910
},
{
"epoch": 2.164090368608799,
"grad_norm": 1.185874011924432,
"learning_rate": 9.024594661705605e-06,
"loss": 1.6543,
"step": 911
},
{
"epoch": 2.1664684898929845,
"grad_norm": 1.3961342001389099,
"learning_rate": 9.005531754865929e-06,
"loss": 1.7178,
"step": 912
},
{
"epoch": 2.16884661117717,
"grad_norm": 1.333944871158627,
"learning_rate": 8.986472497269699e-06,
"loss": 1.701,
"step": 913
},
{
"epoch": 2.1712247324613556,
"grad_norm": 1.6434825073753228,
"learning_rate": 8.967416958855666e-06,
"loss": 1.8314,
"step": 914
},
{
"epoch": 2.173602853745541,
"grad_norm": 1.270123334452636,
"learning_rate": 8.948365209548943e-06,
"loss": 1.8827,
"step": 915
},
{
"epoch": 2.1759809750297263,
"grad_norm": 1.153010176885914,
"learning_rate": 8.929317319260727e-06,
"loss": 1.8395,
"step": 916
},
{
"epoch": 2.178359096313912,
"grad_norm": 1.4178416006595018,
"learning_rate": 8.91027335788806e-06,
"loss": 1.8333,
"step": 917
},
{
"epoch": 2.1807372175980975,
"grad_norm": 1.1553451804083652,
"learning_rate": 8.891233395313572e-06,
"loss": 1.7945,
"step": 918
},
{
"epoch": 2.183115338882283,
"grad_norm": 1.2293490645752978,
"learning_rate": 8.872197501405208e-06,
"loss": 1.7454,
"step": 919
},
{
"epoch": 2.1854934601664686,
"grad_norm": 1.2419979870389453,
"learning_rate": 8.853165746015997e-06,
"loss": 1.8377,
"step": 920
},
{
"epoch": 2.187871581450654,
"grad_norm": 1.7147597799918568,
"learning_rate": 8.834138198983763e-06,
"loss": 1.7252,
"step": 921
},
{
"epoch": 2.1902497027348393,
"grad_norm": 1.2306648279680026,
"learning_rate": 8.815114930130906e-06,
"loss": 1.6057,
"step": 922
},
{
"epoch": 2.192627824019025,
"grad_norm": 1.1859169425356446,
"learning_rate": 8.796096009264114e-06,
"loss": 1.6553,
"step": 923
},
{
"epoch": 2.1950059453032105,
"grad_norm": 1.1602261640700344,
"learning_rate": 8.777081506174127e-06,
"loss": 1.481,
"step": 924
},
{
"epoch": 2.197384066587396,
"grad_norm": 1.402945789419062,
"learning_rate": 8.758071490635468e-06,
"loss": 1.9423,
"step": 925
},
{
"epoch": 2.1997621878715816,
"grad_norm": 1.081930558901166,
"learning_rate": 8.739066032406201e-06,
"loss": 1.5998,
"step": 926
},
{
"epoch": 2.2021403091557668,
"grad_norm": 1.1850370906254002,
"learning_rate": 8.720065201227656e-06,
"loss": 1.4878,
"step": 927
},
{
"epoch": 2.2045184304399523,
"grad_norm": 1.1307191491979665,
"learning_rate": 8.701069066824195e-06,
"loss": 1.6825,
"step": 928
},
{
"epoch": 2.206896551724138,
"grad_norm": 1.2308894920595725,
"learning_rate": 8.68207769890294e-06,
"loss": 1.6438,
"step": 929
},
{
"epoch": 2.2092746730083235,
"grad_norm": 1.3885937358394551,
"learning_rate": 8.663091167153516e-06,
"loss": 1.7407,
"step": 930
},
{
"epoch": 2.211652794292509,
"grad_norm": 1.3493954256475922,
"learning_rate": 8.644109541247811e-06,
"loss": 1.7611,
"step": 931
},
{
"epoch": 2.2140309155766946,
"grad_norm": 1.0683209882742306,
"learning_rate": 8.625132890839706e-06,
"loss": 1.5902,
"step": 932
},
{
"epoch": 2.2164090368608798,
"grad_norm": 1.3428507912542718,
"learning_rate": 8.606161285564826e-06,
"loss": 1.8124,
"step": 933
},
{
"epoch": 2.2187871581450653,
"grad_norm": 1.12745278115296,
"learning_rate": 8.587194795040286e-06,
"loss": 1.7277,
"step": 934
},
{
"epoch": 2.221165279429251,
"grad_norm": 1.214273485118672,
"learning_rate": 8.56823348886442e-06,
"loss": 1.778,
"step": 935
},
{
"epoch": 2.2235434007134365,
"grad_norm": 1.1850085474869128,
"learning_rate": 8.549277436616551e-06,
"loss": 1.6415,
"step": 936
},
{
"epoch": 2.225921521997622,
"grad_norm": 1.0025465720081261,
"learning_rate": 8.530326707856716e-06,
"loss": 1.5837,
"step": 937
},
{
"epoch": 2.228299643281807,
"grad_norm": 1.447315774100193,
"learning_rate": 8.511381372125422e-06,
"loss": 1.7127,
"step": 938
},
{
"epoch": 2.2306777645659928,
"grad_norm": 1.206599818728477,
"learning_rate": 8.492441498943378e-06,
"loss": 1.745,
"step": 939
},
{
"epoch": 2.2330558858501783,
"grad_norm": 1.299850116375868,
"learning_rate": 8.473507157811254e-06,
"loss": 1.645,
"step": 940
},
{
"epoch": 2.235434007134364,
"grad_norm": 1.0985468078995482,
"learning_rate": 8.454578418209421e-06,
"loss": 1.6661,
"step": 941
},
{
"epoch": 2.2378121284185495,
"grad_norm": 1.1275974183027477,
"learning_rate": 8.43565534959769e-06,
"loss": 1.7296,
"step": 942
},
{
"epoch": 2.240190249702735,
"grad_norm": 1.474376638809395,
"learning_rate": 8.416738021415069e-06,
"loss": 1.7114,
"step": 943
},
{
"epoch": 2.24256837098692,
"grad_norm": 1.243674764872066,
"learning_rate": 8.397826503079489e-06,
"loss": 1.7362,
"step": 944
},
{
"epoch": 2.2449464922711058,
"grad_norm": 1.217870809816204,
"learning_rate": 8.378920863987576e-06,
"loss": 1.7632,
"step": 945
},
{
"epoch": 2.2449464922711058,
"eval_loss": 2.4208853244781494,
"eval_runtime": 65.7758,
"eval_samples_per_second": 9.715,
"eval_steps_per_second": 1.216,
"step": 945
},
{
"epoch": 2.2473246135552913,
"grad_norm": 1.1837390067043532,
"learning_rate": 8.360021173514373e-06,
"loss": 1.802,
"step": 946
},
{
"epoch": 2.249702734839477,
"grad_norm": 1.1447223958071244,
"learning_rate": 8.341127501013096e-06,
"loss": 1.8955,
"step": 947
},
{
"epoch": 2.2520808561236625,
"grad_norm": 1.2318620845180448,
"learning_rate": 8.32223991581488e-06,
"loss": 1.7863,
"step": 948
},
{
"epoch": 2.2544589774078476,
"grad_norm": 1.394940898102735,
"learning_rate": 8.30335848722852e-06,
"loss": 1.7229,
"step": 949
},
{
"epoch": 2.256837098692033,
"grad_norm": 1.1301876418286094,
"learning_rate": 8.284483284540217e-06,
"loss": 1.7583,
"step": 950
},
{
"epoch": 2.2592152199762188,
"grad_norm": 1.217726889827069,
"learning_rate": 8.265614377013332e-06,
"loss": 1.5894,
"step": 951
},
{
"epoch": 2.2615933412604043,
"grad_norm": 1.1841467686154767,
"learning_rate": 8.246751833888122e-06,
"loss": 1.6364,
"step": 952
},
{
"epoch": 2.26397146254459,
"grad_norm": 1.1869481085156783,
"learning_rate": 8.22789572438149e-06,
"loss": 1.612,
"step": 953
},
{
"epoch": 2.2663495838287755,
"grad_norm": 1.1712814400466198,
"learning_rate": 8.20904611768673e-06,
"loss": 1.7397,
"step": 954
},
{
"epoch": 2.2687277051129606,
"grad_norm": 1.3442733936079978,
"learning_rate": 8.190203082973272e-06,
"loss": 1.7296,
"step": 955
},
{
"epoch": 2.271105826397146,
"grad_norm": 1.1056459089566495,
"learning_rate": 8.171366689386433e-06,
"loss": 1.6519,
"step": 956
},
{
"epoch": 2.2734839476813318,
"grad_norm": 1.1072199930457924,
"learning_rate": 8.152537006047161e-06,
"loss": 1.6024,
"step": 957
},
{
"epoch": 2.2758620689655173,
"grad_norm": 1.3049195548897443,
"learning_rate": 8.133714102051774e-06,
"loss": 1.825,
"step": 958
},
{
"epoch": 2.278240190249703,
"grad_norm": 1.1216698513231085,
"learning_rate": 8.114898046471721e-06,
"loss": 1.6569,
"step": 959
},
{
"epoch": 2.280618311533888,
"grad_norm": 1.3793482280459277,
"learning_rate": 8.096088908353316e-06,
"loss": 1.7476,
"step": 960
},
{
"epoch": 2.2829964328180736,
"grad_norm": 1.3318314124782873,
"learning_rate": 8.077286756717488e-06,
"loss": 1.7962,
"step": 961
},
{
"epoch": 2.285374554102259,
"grad_norm": 1.165705372513211,
"learning_rate": 8.058491660559536e-06,
"loss": 1.708,
"step": 962
},
{
"epoch": 2.287752675386445,
"grad_norm": 1.0416513999223338,
"learning_rate": 8.039703688848858e-06,
"loss": 1.6261,
"step": 963
},
{
"epoch": 2.2901307966706304,
"grad_norm": 1.2426839446823297,
"learning_rate": 8.020922910528717e-06,
"loss": 1.7174,
"step": 964
},
{
"epoch": 2.292508917954816,
"grad_norm": 1.2325341409873833,
"learning_rate": 8.002149394515973e-06,
"loss": 1.4361,
"step": 965
},
{
"epoch": 2.294887039239001,
"grad_norm": 1.055835642253928,
"learning_rate": 7.983383209700839e-06,
"loss": 1.642,
"step": 966
},
{
"epoch": 2.2972651605231866,
"grad_norm": 1.1036311323131265,
"learning_rate": 7.96462442494663e-06,
"loss": 1.5686,
"step": 967
},
{
"epoch": 2.299643281807372,
"grad_norm": 1.165719367931869,
"learning_rate": 7.945873109089503e-06,
"loss": 1.6485,
"step": 968
},
{
"epoch": 2.302021403091558,
"grad_norm": 1.292263443675501,
"learning_rate": 7.92712933093821e-06,
"loss": 1.4937,
"step": 969
},
{
"epoch": 2.304399524375743,
"grad_norm": 1.2446616344126429,
"learning_rate": 7.908393159273835e-06,
"loss": 1.61,
"step": 970
},
{
"epoch": 2.3067776456599285,
"grad_norm": 1.2958114102555907,
"learning_rate": 7.88966466284956e-06,
"loss": 1.7275,
"step": 971
},
{
"epoch": 2.309155766944114,
"grad_norm": 1.197050064334699,
"learning_rate": 7.870943910390392e-06,
"loss": 1.6454,
"step": 972
},
{
"epoch": 2.3115338882282996,
"grad_norm": 1.2307123050224071,
"learning_rate": 7.852230970592936e-06,
"loss": 1.6779,
"step": 973
},
{
"epoch": 2.313912009512485,
"grad_norm": 1.0920081628895666,
"learning_rate": 7.83352591212511e-06,
"loss": 1.6522,
"step": 974
},
{
"epoch": 2.316290130796671,
"grad_norm": 1.1702109811198733,
"learning_rate": 7.814828803625926e-06,
"loss": 1.5587,
"step": 975
},
{
"epoch": 2.3186682520808564,
"grad_norm": 1.4321131869810775,
"learning_rate": 7.796139713705214e-06,
"loss": 2.0141,
"step": 976
},
{
"epoch": 2.3210463733650415,
"grad_norm": 1.175162275052099,
"learning_rate": 7.777458710943384e-06,
"loss": 1.6676,
"step": 977
},
{
"epoch": 2.323424494649227,
"grad_norm": 1.286998127575608,
"learning_rate": 7.758785863891171e-06,
"loss": 1.7923,
"step": 978
},
{
"epoch": 2.3258026159334126,
"grad_norm": 1.2418872677949764,
"learning_rate": 7.740121241069376e-06,
"loss": 1.6202,
"step": 979
},
{
"epoch": 2.328180737217598,
"grad_norm": 1.2186583834505373,
"learning_rate": 7.721464910968628e-06,
"loss": 1.5604,
"step": 980
},
{
"epoch": 2.3305588585017833,
"grad_norm": 1.223129867370841,
"learning_rate": 7.702816942049118e-06,
"loss": 1.5994,
"step": 981
},
{
"epoch": 2.332936979785969,
"grad_norm": 1.1668829594845642,
"learning_rate": 7.684177402740365e-06,
"loss": 1.7446,
"step": 982
},
{
"epoch": 2.3353151010701545,
"grad_norm": 1.227131128900792,
"learning_rate": 7.66554636144095e-06,
"loss": 1.6967,
"step": 983
},
{
"epoch": 2.33769322235434,
"grad_norm": 1.1298937223507048,
"learning_rate": 7.646923886518263e-06,
"loss": 1.6147,
"step": 984
},
{
"epoch": 2.3400713436385256,
"grad_norm": 1.1172424007795727,
"learning_rate": 7.628310046308272e-06,
"loss": 1.8183,
"step": 985
},
{
"epoch": 2.342449464922711,
"grad_norm": 1.0995723009825376,
"learning_rate": 7.609704909115254e-06,
"loss": 1.7295,
"step": 986
},
{
"epoch": 2.344827586206897,
"grad_norm": 2.166347270097445,
"learning_rate": 7.5911085432115476e-06,
"loss": 1.8549,
"step": 987
},
{
"epoch": 2.347205707491082,
"grad_norm": 1.225917922436242,
"learning_rate": 7.57252101683731e-06,
"loss": 1.392,
"step": 988
},
{
"epoch": 2.3495838287752675,
"grad_norm": 1.1738557938320278,
"learning_rate": 7.5539423982002535e-06,
"loss": 1.6863,
"step": 989
},
{
"epoch": 2.351961950059453,
"grad_norm": 1.1432224564501046,
"learning_rate": 7.535372755475411e-06,
"loss": 1.7436,
"step": 990
},
{
"epoch": 2.3543400713436387,
"grad_norm": 1.1640247664089125,
"learning_rate": 7.516812156804874e-06,
"loss": 1.7646,
"step": 991
},
{
"epoch": 2.356718192627824,
"grad_norm": 1.2793531458628074,
"learning_rate": 7.4982606702975505e-06,
"loss": 1.6448,
"step": 992
},
{
"epoch": 2.3590963139120094,
"grad_norm": 1.1162056385630008,
"learning_rate": 7.479718364028903e-06,
"loss": 1.7002,
"step": 993
},
{
"epoch": 2.361474435196195,
"grad_norm": 1.1444550748454578,
"learning_rate": 7.461185306040712e-06,
"loss": 1.7143,
"step": 994
},
{
"epoch": 2.3638525564803805,
"grad_norm": 1.1918354987814666,
"learning_rate": 7.442661564340823e-06,
"loss": 1.5962,
"step": 995
},
{
"epoch": 2.366230677764566,
"grad_norm": 1.2320281238709774,
"learning_rate": 7.4241472069028915e-06,
"loss": 1.8181,
"step": 996
},
{
"epoch": 2.3686087990487517,
"grad_norm": 1.2762221347684297,
"learning_rate": 7.4056423016661405e-06,
"loss": 1.765,
"step": 997
},
{
"epoch": 2.370986920332937,
"grad_norm": 1.1579306898032722,
"learning_rate": 7.3871469165351015e-06,
"loss": 1.5342,
"step": 998
},
{
"epoch": 2.3733650416171224,
"grad_norm": 1.234792249302144,
"learning_rate": 7.368661119379378e-06,
"loss": 1.8651,
"step": 999
},
{
"epoch": 2.375743162901308,
"grad_norm": 1.3917341321909777,
"learning_rate": 7.350184978033386e-06,
"loss": 1.782,
"step": 1000
},
{
"epoch": 2.3781212841854935,
"grad_norm": 1.29943881237625,
"learning_rate": 7.331718560296109e-06,
"loss": 1.615,
"step": 1001
},
{
"epoch": 2.380499405469679,
"grad_norm": 1.3084311146501468,
"learning_rate": 7.313261933930858e-06,
"loss": 1.7772,
"step": 1002
},
{
"epoch": 2.382877526753864,
"grad_norm": 1.162552758386455,
"learning_rate": 7.294815166664998e-06,
"loss": 1.8071,
"step": 1003
},
{
"epoch": 2.38525564803805,
"grad_norm": 1.1614300990893485,
"learning_rate": 7.276378326189729e-06,
"loss": 1.516,
"step": 1004
},
{
"epoch": 2.3876337693222354,
"grad_norm": 1.154540065487417,
"learning_rate": 7.25795148015982e-06,
"loss": 1.604,
"step": 1005
},
{
"epoch": 2.390011890606421,
"grad_norm": 1.2082947661643852,
"learning_rate": 7.2395346961933645e-06,
"loss": 1.5811,
"step": 1006
},
{
"epoch": 2.3923900118906065,
"grad_norm": 1.2336019557768343,
"learning_rate": 7.221128041871532e-06,
"loss": 1.7729,
"step": 1007
},
{
"epoch": 2.394768133174792,
"grad_norm": 1.1152227970531388,
"learning_rate": 7.202731584738323e-06,
"loss": 1.7409,
"step": 1008
},
{
"epoch": 2.397146254458977,
"grad_norm": 1.3490125025605684,
"learning_rate": 7.184345392300318e-06,
"loss": 1.6814,
"step": 1009
},
{
"epoch": 2.399524375743163,
"grad_norm": 1.1548182248428795,
"learning_rate": 7.16596953202643e-06,
"loss": 1.6186,
"step": 1010
},
{
"epoch": 2.4019024970273484,
"grad_norm": 1.2852743643592273,
"learning_rate": 7.1476040713476605e-06,
"loss": 1.8935,
"step": 1011
},
{
"epoch": 2.404280618311534,
"grad_norm": 1.220532077802057,
"learning_rate": 7.129249077656844e-06,
"loss": 1.6207,
"step": 1012
},
{
"epoch": 2.4066587395957195,
"grad_norm": 1.2402232079606048,
"learning_rate": 7.110904618308409e-06,
"loss": 1.8316,
"step": 1013
},
{
"epoch": 2.4090368608799047,
"grad_norm": 1.1273649207025187,
"learning_rate": 7.092570760618132e-06,
"loss": 1.7531,
"step": 1014
},
{
"epoch": 2.4114149821640902,
"grad_norm": 1.1542485461767478,
"learning_rate": 7.074247571862877e-06,
"loss": 1.4423,
"step": 1015
},
{
"epoch": 2.413793103448276,
"grad_norm": 1.08202605516453,
"learning_rate": 7.055935119280369e-06,
"loss": 1.6266,
"step": 1016
},
{
"epoch": 2.4161712247324614,
"grad_norm": 1.1770704752495105,
"learning_rate": 7.0376334700689254e-06,
"loss": 1.6228,
"step": 1017
},
{
"epoch": 2.418549346016647,
"grad_norm": 1.456786984171917,
"learning_rate": 7.019342691387225e-06,
"loss": 1.7052,
"step": 1018
},
{
"epoch": 2.4209274673008325,
"grad_norm": 1.1838036075763612,
"learning_rate": 7.001062850354059e-06,
"loss": 1.7753,
"step": 1019
},
{
"epoch": 2.4233055885850177,
"grad_norm": 1.1306005750393724,
"learning_rate": 6.9827940140480776e-06,
"loss": 1.67,
"step": 1020
},
{
"epoch": 2.4256837098692032,
"grad_norm": 1.3893043397375318,
"learning_rate": 6.964536249507556e-06,
"loss": 1.9198,
"step": 1021
},
{
"epoch": 2.428061831153389,
"grad_norm": 1.1678798868137186,
"learning_rate": 6.946289623730131e-06,
"loss": 1.6164,
"step": 1022
},
{
"epoch": 2.4304399524375744,
"grad_norm": 1.154762934924946,
"learning_rate": 6.928054203672572e-06,
"loss": 1.7154,
"step": 1023
},
{
"epoch": 2.43281807372176,
"grad_norm": 1.1232082829911434,
"learning_rate": 6.909830056250527e-06,
"loss": 1.7256,
"step": 1024
},
{
"epoch": 2.435196195005945,
"grad_norm": 1.196133461314455,
"learning_rate": 6.891617248338282e-06,
"loss": 1.8479,
"step": 1025
},
{
"epoch": 2.4375743162901307,
"grad_norm": 1.2594174079784615,
"learning_rate": 6.873415846768503e-06,
"loss": 1.7605,
"step": 1026
},
{
"epoch": 2.4399524375743162,
"grad_norm": 1.1659462826012594,
"learning_rate": 6.85522591833201e-06,
"loss": 1.8646,
"step": 1027
},
{
"epoch": 2.442330558858502,
"grad_norm": 1.2914773691659898,
"learning_rate": 6.837047529777516e-06,
"loss": 1.5371,
"step": 1028
},
{
"epoch": 2.4447086801426874,
"grad_norm": 1.0754377635014698,
"learning_rate": 6.8188807478113904e-06,
"loss": 1.6698,
"step": 1029
},
{
"epoch": 2.447086801426873,
"grad_norm": 1.113563098766549,
"learning_rate": 6.800725639097412e-06,
"loss": 1.3603,
"step": 1030
},
{
"epoch": 2.449464922711058,
"grad_norm": 1.128889361821313,
"learning_rate": 6.782582270256519e-06,
"loss": 1.5279,
"step": 1031
},
{
"epoch": 2.4518430439952437,
"grad_norm": 1.329815666398893,
"learning_rate": 6.764450707866577e-06,
"loss": 1.8093,
"step": 1032
},
{
"epoch": 2.4542211652794292,
"grad_norm": 1.3107218996988408,
"learning_rate": 6.746331018462122e-06,
"loss": 1.6953,
"step": 1033
},
{
"epoch": 2.456599286563615,
"grad_norm": 1.3794075786425875,
"learning_rate": 6.728223268534127e-06,
"loss": 1.5287,
"step": 1034
},
{
"epoch": 2.4589774078478004,
"grad_norm": 1.2022506712726995,
"learning_rate": 6.710127524529746e-06,
"loss": 1.7323,
"step": 1035
},
{
"epoch": 2.4613555291319855,
"grad_norm": 1.2379907256486176,
"learning_rate": 6.6920438528520794e-06,
"loss": 1.7864,
"step": 1036
},
{
"epoch": 2.463733650416171,
"grad_norm": 1.2398631324380247,
"learning_rate": 6.673972319859928e-06,
"loss": 1.6605,
"step": 1037
},
{
"epoch": 2.4661117717003567,
"grad_norm": 1.4244397199332064,
"learning_rate": 6.655912991867551e-06,
"loss": 1.5941,
"step": 1038
},
{
"epoch": 2.4684898929845422,
"grad_norm": 1.0907158228131357,
"learning_rate": 6.6378659351444185e-06,
"loss": 1.6619,
"step": 1039
},
{
"epoch": 2.470868014268728,
"grad_norm": 1.2326470273897912,
"learning_rate": 6.619831215914974e-06,
"loss": 1.5241,
"step": 1040
},
{
"epoch": 2.4732461355529134,
"grad_norm": 1.1815306018044767,
"learning_rate": 6.601808900358382e-06,
"loss": 1.5294,
"step": 1041
},
{
"epoch": 2.4756242568370985,
"grad_norm": 1.3514710702412487,
"learning_rate": 6.583799054608293e-06,
"loss": 1.7267,
"step": 1042
},
{
"epoch": 2.478002378121284,
"grad_norm": 1.121258311049222,
"learning_rate": 6.565801744752604e-06,
"loss": 1.8033,
"step": 1043
},
{
"epoch": 2.4803804994054697,
"grad_norm": 1.0430171184122976,
"learning_rate": 6.547817036833208e-06,
"loss": 1.6306,
"step": 1044
},
{
"epoch": 2.4827586206896552,
"grad_norm": 1.1509719987044542,
"learning_rate": 6.529844996845751e-06,
"loss": 1.7799,
"step": 1045
},
{
"epoch": 2.485136741973841,
"grad_norm": 1.3436366916936602,
"learning_rate": 6.511885690739399e-06,
"loss": 1.7702,
"step": 1046
},
{
"epoch": 2.487514863258026,
"grad_norm": 1.2980206283561453,
"learning_rate": 6.4939391844165865e-06,
"loss": 1.6951,
"step": 1047
},
{
"epoch": 2.4898929845422115,
"grad_norm": 1.2496756975636276,
"learning_rate": 6.476005543732783e-06,
"loss": 1.6498,
"step": 1048
},
{
"epoch": 2.492271105826397,
"grad_norm": 1.5028507878209147,
"learning_rate": 6.4580848344962435e-06,
"loss": 1.6611,
"step": 1049
},
{
"epoch": 2.4946492271105827,
"grad_norm": 1.2800840762788763,
"learning_rate": 6.440177122467769e-06,
"loss": 1.8898,
"step": 1050
},
{
"epoch": 2.4946492271105827,
"eval_loss": 2.416064977645874,
"eval_runtime": 66.0363,
"eval_samples_per_second": 9.676,
"eval_steps_per_second": 1.211,
"step": 1050
},
{
"epoch": 2.4970273483947683,
"grad_norm": 1.1873534107751536,
"learning_rate": 6.422282473360471e-06,
"loss": 1.6971,
"step": 1051
},
{
"epoch": 2.499405469678954,
"grad_norm": 1.1655553131443337,
"learning_rate": 6.404400952839522e-06,
"loss": 1.6465,
"step": 1052
},
{
"epoch": 2.501783590963139,
"grad_norm": 1.3302997667501615,
"learning_rate": 6.386532626521924e-06,
"loss": 1.5254,
"step": 1053
},
{
"epoch": 2.5041617122473245,
"grad_norm": 1.0711049335495844,
"learning_rate": 6.36867755997626e-06,
"loss": 1.674,
"step": 1054
},
{
"epoch": 2.50653983353151,
"grad_norm": 1.1545823396692283,
"learning_rate": 6.350835818722449e-06,
"loss": 1.5929,
"step": 1055
},
{
"epoch": 2.5089179548156957,
"grad_norm": 1.329273500711881,
"learning_rate": 6.333007468231521e-06,
"loss": 1.8369,
"step": 1056
},
{
"epoch": 2.5112960760998813,
"grad_norm": 1.1793439619345145,
"learning_rate": 6.315192573925366e-06,
"loss": 1.8477,
"step": 1057
},
{
"epoch": 2.5136741973840664,
"grad_norm": 1.291888497964125,
"learning_rate": 6.297391201176495e-06,
"loss": 1.7726,
"step": 1058
},
{
"epoch": 2.516052318668252,
"grad_norm": 1.4228282247113229,
"learning_rate": 6.2796034153077976e-06,
"loss": 1.8691,
"step": 1059
},
{
"epoch": 2.5184304399524375,
"grad_norm": 1.1530624413295318,
"learning_rate": 6.261829281592313e-06,
"loss": 1.7016,
"step": 1060
},
{
"epoch": 2.520808561236623,
"grad_norm": 1.1299743740921777,
"learning_rate": 6.244068865252979e-06,
"loss": 1.599,
"step": 1061
},
{
"epoch": 2.5231866825208087,
"grad_norm": 1.626142124032786,
"learning_rate": 6.226322231462394e-06,
"loss": 2.0729,
"step": 1062
},
{
"epoch": 2.5255648038049943,
"grad_norm": 1.2190327254010218,
"learning_rate": 6.20858944534259e-06,
"loss": 1.6418,
"step": 1063
},
{
"epoch": 2.5279429250891794,
"grad_norm": 1.4103004422055179,
"learning_rate": 6.1908705719647735e-06,
"loss": 1.6974,
"step": 1064
},
{
"epoch": 2.530321046373365,
"grad_norm": 1.2555931729554382,
"learning_rate": 6.173165676349103e-06,
"loss": 1.7555,
"step": 1065
},
{
"epoch": 2.5326991676575505,
"grad_norm": 1.3278788007077313,
"learning_rate": 6.155474823464446e-06,
"loss": 1.8747,
"step": 1066
},
{
"epoch": 2.535077288941736,
"grad_norm": 1.0206907165277292,
"learning_rate": 6.137798078228139e-06,
"loss": 1.7062,
"step": 1067
},
{
"epoch": 2.5374554102259212,
"grad_norm": 1.1678494099359702,
"learning_rate": 6.1201355055057486e-06,
"loss": 1.6618,
"step": 1068
},
{
"epoch": 2.539833531510107,
"grad_norm": 1.276758941048691,
"learning_rate": 6.1024871701108345e-06,
"loss": 1.5699,
"step": 1069
},
{
"epoch": 2.5422116527942924,
"grad_norm": 1.4965690225241113,
"learning_rate": 6.084853136804711e-06,
"loss": 1.8962,
"step": 1070
},
{
"epoch": 2.544589774078478,
"grad_norm": 1.2781827417360803,
"learning_rate": 6.067233470296216e-06,
"loss": 1.6127,
"step": 1071
},
{
"epoch": 2.5469678953626635,
"grad_norm": 1.3693753990851296,
"learning_rate": 6.049628235241459e-06,
"loss": 1.7889,
"step": 1072
},
{
"epoch": 2.549346016646849,
"grad_norm": 1.228145174812121,
"learning_rate": 6.032037496243604e-06,
"loss": 1.5608,
"step": 1073
},
{
"epoch": 2.5517241379310347,
"grad_norm": 1.2063099794472876,
"learning_rate": 6.014461317852606e-06,
"loss": 1.7348,
"step": 1074
},
{
"epoch": 2.55410225921522,
"grad_norm": 1.217042589994472,
"learning_rate": 5.996899764565005e-06,
"loss": 1.6992,
"step": 1075
},
{
"epoch": 2.5564803804994054,
"grad_norm": 1.301014047396616,
"learning_rate": 5.979352900823663e-06,
"loss": 1.8162,
"step": 1076
},
{
"epoch": 2.558858501783591,
"grad_norm": 1.0567226366943008,
"learning_rate": 5.961820791017544e-06,
"loss": 1.6215,
"step": 1077
},
{
"epoch": 2.5612366230677766,
"grad_norm": 1.1919768727324513,
"learning_rate": 5.9443034994814655e-06,
"loss": 1.7082,
"step": 1078
},
{
"epoch": 2.5636147443519617,
"grad_norm": 1.116399660030849,
"learning_rate": 5.926801090495873e-06,
"loss": 1.7973,
"step": 1079
},
{
"epoch": 2.5659928656361473,
"grad_norm": 2.0791551718173236,
"learning_rate": 5.9093136282866014e-06,
"loss": 1.593,
"step": 1080
},
{
"epoch": 2.568370986920333,
"grad_norm": 1.102934034890405,
"learning_rate": 5.891841177024635e-06,
"loss": 1.7117,
"step": 1081
},
{
"epoch": 2.5707491082045184,
"grad_norm": 1.4346461019869596,
"learning_rate": 5.874383800825878e-06,
"loss": 1.7807,
"step": 1082
},
{
"epoch": 2.573127229488704,
"grad_norm": 1.3171902324738824,
"learning_rate": 5.856941563750907e-06,
"loss": 1.7296,
"step": 1083
},
{
"epoch": 2.5755053507728896,
"grad_norm": 1.3868744564812188,
"learning_rate": 5.839514529804757e-06,
"loss": 1.84,
"step": 1084
},
{
"epoch": 2.577883472057075,
"grad_norm": 1.2362063868655484,
"learning_rate": 5.822102762936667e-06,
"loss": 1.4915,
"step": 1085
},
{
"epoch": 2.5802615933412603,
"grad_norm": 1.2404289042620804,
"learning_rate": 5.8047063270398566e-06,
"loss": 1.6799,
"step": 1086
},
{
"epoch": 2.582639714625446,
"grad_norm": 1.1129231989370107,
"learning_rate": 5.78732528595128e-06,
"loss": 1.7227,
"step": 1087
},
{
"epoch": 2.5850178359096314,
"grad_norm": 1.161432165715355,
"learning_rate": 5.769959703451413e-06,
"loss": 1.7592,
"step": 1088
},
{
"epoch": 2.587395957193817,
"grad_norm": 1.248731815613165,
"learning_rate": 5.752609643263996e-06,
"loss": 1.5589,
"step": 1089
},
{
"epoch": 2.589774078478002,
"grad_norm": 1.1857883701158243,
"learning_rate": 5.7352751690558025e-06,
"loss": 1.8265,
"step": 1090
},
{
"epoch": 2.5921521997621877,
"grad_norm": 1.2318137626074095,
"learning_rate": 5.717956344436435e-06,
"loss": 1.8188,
"step": 1091
},
{
"epoch": 2.5945303210463733,
"grad_norm": 1.2118351181535407,
"learning_rate": 5.700653232958047e-06,
"loss": 1.6086,
"step": 1092
},
{
"epoch": 2.596908442330559,
"grad_norm": 1.3511727528093584,
"learning_rate": 5.683365898115145e-06,
"loss": 1.8693,
"step": 1093
},
{
"epoch": 2.5992865636147444,
"grad_norm": 1.0915962232836285,
"learning_rate": 5.666094403344336e-06,
"loss": 1.6513,
"step": 1094
},
{
"epoch": 2.60166468489893,
"grad_norm": 1.1047451691863088,
"learning_rate": 5.6488388120241e-06,
"loss": 1.5876,
"step": 1095
},
{
"epoch": 2.6040428061831156,
"grad_norm": 1.0955354456142385,
"learning_rate": 5.63159918747457e-06,
"loss": 1.7715,
"step": 1096
},
{
"epoch": 2.6064209274673007,
"grad_norm": 1.4037474761385684,
"learning_rate": 5.6143755929572704e-06,
"loss": 1.9165,
"step": 1097
},
{
"epoch": 2.6087990487514863,
"grad_norm": 1.210647433502284,
"learning_rate": 5.5971680916749226e-06,
"loss": 1.6276,
"step": 1098
},
{
"epoch": 2.611177170035672,
"grad_norm": 1.3342524117099221,
"learning_rate": 5.579976746771179e-06,
"loss": 1.874,
"step": 1099
},
{
"epoch": 2.6135552913198574,
"grad_norm": 1.219886735930388,
"learning_rate": 5.562801621330402e-06,
"loss": 1.6268,
"step": 1100
},
{
"epoch": 2.6159334126040426,
"grad_norm": 1.254801496767963,
"learning_rate": 5.545642778377457e-06,
"loss": 1.7502,
"step": 1101
},
{
"epoch": 2.618311533888228,
"grad_norm": 1.1841203502125026,
"learning_rate": 5.5285002808774345e-06,
"loss": 1.8315,
"step": 1102
},
{
"epoch": 2.6206896551724137,
"grad_norm": 1.4994607858303892,
"learning_rate": 5.511374191735467e-06,
"loss": 1.8898,
"step": 1103
},
{
"epoch": 2.6230677764565993,
"grad_norm": 1.303031931946195,
"learning_rate": 5.49426457379646e-06,
"loss": 1.5813,
"step": 1104
},
{
"epoch": 2.625445897740785,
"grad_norm": 1.1926132163625953,
"learning_rate": 5.4771714898448816e-06,
"loss": 1.5727,
"step": 1105
},
{
"epoch": 2.6278240190249704,
"grad_norm": 1.2672315312487659,
"learning_rate": 5.460095002604533e-06,
"loss": 1.6712,
"step": 1106
},
{
"epoch": 2.630202140309156,
"grad_norm": 1.2855515231934755,
"learning_rate": 5.4430351747383045e-06,
"loss": 1.5207,
"step": 1107
},
{
"epoch": 2.632580261593341,
"grad_norm": 1.262934558418249,
"learning_rate": 5.425992068847965e-06,
"loss": 1.5892,
"step": 1108
},
{
"epoch": 2.6349583828775267,
"grad_norm": 1.2176816137166029,
"learning_rate": 5.408965747473913e-06,
"loss": 1.7342,
"step": 1109
},
{
"epoch": 2.6373365041617123,
"grad_norm": 1.2492689217919521,
"learning_rate": 5.391956273094952e-06,
"loss": 1.6075,
"step": 1110
},
{
"epoch": 2.639714625445898,
"grad_norm": 1.0855880724356506,
"learning_rate": 5.374963708128076e-06,
"loss": 1.6012,
"step": 1111
},
{
"epoch": 2.642092746730083,
"grad_norm": 1.092980654609454,
"learning_rate": 5.357988114928221e-06,
"loss": 1.79,
"step": 1112
},
{
"epoch": 2.6444708680142686,
"grad_norm": 1.214623091364747,
"learning_rate": 5.341029555788051e-06,
"loss": 1.6981,
"step": 1113
},
{
"epoch": 2.646848989298454,
"grad_norm": 1.3184091992263423,
"learning_rate": 5.324088092937716e-06,
"loss": 1.6904,
"step": 1114
},
{
"epoch": 2.6492271105826397,
"grad_norm": 1.1971392521428008,
"learning_rate": 5.30716378854463e-06,
"loss": 1.698,
"step": 1115
},
{
"epoch": 2.6516052318668253,
"grad_norm": 1.0548412087042252,
"learning_rate": 5.2902567047132505e-06,
"loss": 1.7051,
"step": 1116
},
{
"epoch": 2.653983353151011,
"grad_norm": 1.2246037817923938,
"learning_rate": 5.27336690348484e-06,
"loss": 1.6302,
"step": 1117
},
{
"epoch": 2.6563614744351964,
"grad_norm": 1.1580499998747145,
"learning_rate": 5.2564944468372344e-06,
"loss": 1.4743,
"step": 1118
},
{
"epoch": 2.6587395957193816,
"grad_norm": 1.0651264431180674,
"learning_rate": 5.239639396684641e-06,
"loss": 1.6479,
"step": 1119
},
{
"epoch": 2.661117717003567,
"grad_norm": 1.243044606481579,
"learning_rate": 5.22280181487737e-06,
"loss": 1.6342,
"step": 1120
},
{
"epoch": 2.6634958382877527,
"grad_norm": 1.278385211653831,
"learning_rate": 5.205981763201653e-06,
"loss": 1.4981,
"step": 1121
},
{
"epoch": 2.6658739595719383,
"grad_norm": 1.3141585309901718,
"learning_rate": 5.189179303379377e-06,
"loss": 1.8093,
"step": 1122
},
{
"epoch": 2.6682520808561234,
"grad_norm": 1.132203960933878,
"learning_rate": 5.172394497067881e-06,
"loss": 1.56,
"step": 1123
},
{
"epoch": 2.670630202140309,
"grad_norm": 1.1928572353153866,
"learning_rate": 5.155627405859731e-06,
"loss": 1.69,
"step": 1124
},
{
"epoch": 2.6730083234244946,
"grad_norm": 1.1862742147972845,
"learning_rate": 5.138878091282472e-06,
"loss": 1.7242,
"step": 1125
},
{
"epoch": 2.67538644470868,
"grad_norm": 1.1450522807579773,
"learning_rate": 5.122146614798433e-06,
"loss": 1.5918,
"step": 1126
},
{
"epoch": 2.6777645659928657,
"grad_norm": 1.1890028639136556,
"learning_rate": 5.105433037804473e-06,
"loss": 1.804,
"step": 1127
},
{
"epoch": 2.6801426872770513,
"grad_norm": 1.228689756576672,
"learning_rate": 5.088737421631767e-06,
"loss": 1.7795,
"step": 1128
},
{
"epoch": 2.682520808561237,
"grad_norm": 1.1598278737067984,
"learning_rate": 5.072059827545598e-06,
"loss": 1.7347,
"step": 1129
},
{
"epoch": 2.684898929845422,
"grad_norm": 1.1998086711545068,
"learning_rate": 5.055400316745096e-06,
"loss": 1.8378,
"step": 1130
},
{
"epoch": 2.6872770511296076,
"grad_norm": 1.13047712313628,
"learning_rate": 5.038758950363052e-06,
"loss": 1.6083,
"step": 1131
},
{
"epoch": 2.689655172413793,
"grad_norm": 1.2433884758695508,
"learning_rate": 5.0221357894656605e-06,
"loss": 1.5978,
"step": 1132
},
{
"epoch": 2.6920332936979787,
"grad_norm": 1.1378408221633547,
"learning_rate": 5.005530895052315e-06,
"loss": 1.7166,
"step": 1133
},
{
"epoch": 2.694411414982164,
"grad_norm": 1.2328573609700277,
"learning_rate": 4.988944328055386e-06,
"loss": 1.5597,
"step": 1134
},
{
"epoch": 2.6967895362663494,
"grad_norm": 1.2760668630843608,
"learning_rate": 4.972376149339978e-06,
"loss": 1.8548,
"step": 1135
},
{
"epoch": 2.699167657550535,
"grad_norm": 1.1500316126915893,
"learning_rate": 4.955826419703736e-06,
"loss": 1.8115,
"step": 1136
},
{
"epoch": 2.7015457788347206,
"grad_norm": 1.1979018516552924,
"learning_rate": 4.939295199876588e-06,
"loss": 1.6296,
"step": 1137
},
{
"epoch": 2.703923900118906,
"grad_norm": 1.2523452493405105,
"learning_rate": 4.922782550520545e-06,
"loss": 1.5373,
"step": 1138
},
{
"epoch": 2.7063020214030917,
"grad_norm": 1.105439461148688,
"learning_rate": 4.906288532229483e-06,
"loss": 1.6174,
"step": 1139
},
{
"epoch": 2.7086801426872773,
"grad_norm": 1.110759656877746,
"learning_rate": 4.889813205528895e-06,
"loss": 1.4889,
"step": 1140
},
{
"epoch": 2.7110582639714624,
"grad_norm": 1.1621478050546636,
"learning_rate": 4.873356630875698e-06,
"loss": 1.7118,
"step": 1141
},
{
"epoch": 2.713436385255648,
"grad_norm": 1.1168812194625557,
"learning_rate": 4.856918868657987e-06,
"loss": 1.7404,
"step": 1142
},
{
"epoch": 2.7158145065398336,
"grad_norm": 1.2085647298687852,
"learning_rate": 4.840499979194827e-06,
"loss": 1.5431,
"step": 1143
},
{
"epoch": 2.718192627824019,
"grad_norm": 1.3206971686698805,
"learning_rate": 4.824100022736036e-06,
"loss": 1.646,
"step": 1144
},
{
"epoch": 2.7205707491082043,
"grad_norm": 1.2005611693463547,
"learning_rate": 4.8077190594619425e-06,
"loss": 1.6168,
"step": 1145
},
{
"epoch": 2.72294887039239,
"grad_norm": 1.3811064744767858,
"learning_rate": 4.791357149483195e-06,
"loss": 1.9062,
"step": 1146
},
{
"epoch": 2.7253269916765754,
"grad_norm": 1.2484162904758584,
"learning_rate": 4.775014352840512e-06,
"loss": 1.8511,
"step": 1147
},
{
"epoch": 2.727705112960761,
"grad_norm": 1.2749931557296785,
"learning_rate": 4.758690729504478e-06,
"loss": 1.7171,
"step": 1148
},
{
"epoch": 2.7300832342449466,
"grad_norm": 1.1889630755038343,
"learning_rate": 4.7423863393753264e-06,
"loss": 1.7778,
"step": 1149
},
{
"epoch": 2.732461355529132,
"grad_norm": 1.4063711129472924,
"learning_rate": 4.726101242282708e-06,
"loss": 1.7224,
"step": 1150
},
{
"epoch": 2.7348394768133177,
"grad_norm": 1.2815162496010937,
"learning_rate": 4.709835497985472e-06,
"loss": 1.663,
"step": 1151
},
{
"epoch": 2.737217598097503,
"grad_norm": 1.1388809085271518,
"learning_rate": 4.693589166171466e-06,
"loss": 1.7866,
"step": 1152
},
{
"epoch": 2.7395957193816884,
"grad_norm": 1.2040063307183857,
"learning_rate": 4.677362306457288e-06,
"loss": 1.4829,
"step": 1153
},
{
"epoch": 2.741973840665874,
"grad_norm": 1.2192242287644692,
"learning_rate": 4.661154978388094e-06,
"loss": 1.5552,
"step": 1154
},
{
"epoch": 2.7443519619500596,
"grad_norm": 1.2655433571109214,
"learning_rate": 4.64496724143736e-06,
"loss": 1.7095,
"step": 1155
},
{
"epoch": 2.7443519619500596,
"eval_loss": 2.399907350540161,
"eval_runtime": 65.5929,
"eval_samples_per_second": 9.742,
"eval_steps_per_second": 1.22,
"step": 1155
},
{
"epoch": 2.7467300832342447,
"grad_norm": 1.179732883618249,
"learning_rate": 4.628799155006669e-06,
"loss": 1.6908,
"step": 1156
},
{
"epoch": 2.7491082045184303,
"grad_norm": 1.3989245145229374,
"learning_rate": 4.61265077842551e-06,
"loss": 1.8142,
"step": 1157
},
{
"epoch": 2.751486325802616,
"grad_norm": 1.1191787120080152,
"learning_rate": 4.596522170951028e-06,
"loss": 1.7384,
"step": 1158
},
{
"epoch": 2.7538644470868014,
"grad_norm": 1.2855659443263352,
"learning_rate": 4.580413391767838e-06,
"loss": 1.9503,
"step": 1159
},
{
"epoch": 2.756242568370987,
"grad_norm": 1.2264542470180242,
"learning_rate": 4.56432449998779e-06,
"loss": 1.7279,
"step": 1160
},
{
"epoch": 2.7586206896551726,
"grad_norm": 1.2812883169114553,
"learning_rate": 4.548255554649748e-06,
"loss": 1.6427,
"step": 1161
},
{
"epoch": 2.760998810939358,
"grad_norm": 1.1781486757390418,
"learning_rate": 4.532206614719398e-06,
"loss": 1.5545,
"step": 1162
},
{
"epoch": 2.7633769322235433,
"grad_norm": 1.2277775509340916,
"learning_rate": 4.5161777390889985e-06,
"loss": 1.7287,
"step": 1163
},
{
"epoch": 2.765755053507729,
"grad_norm": 1.1479936742945047,
"learning_rate": 4.500168986577197e-06,
"loss": 1.6304,
"step": 1164
},
{
"epoch": 2.7681331747919145,
"grad_norm": 1.1624832811933103,
"learning_rate": 4.484180415928786e-06,
"loss": 1.4935,
"step": 1165
},
{
"epoch": 2.7705112960761,
"grad_norm": 1.3218291013979937,
"learning_rate": 4.468212085814501e-06,
"loss": 1.7302,
"step": 1166
},
{
"epoch": 2.772889417360285,
"grad_norm": 1.1796953319671484,
"learning_rate": 4.452264054830815e-06,
"loss": 1.7666,
"step": 1167
},
{
"epoch": 2.7752675386444707,
"grad_norm": 1.0832095485324484,
"learning_rate": 4.436336381499701e-06,
"loss": 1.6282,
"step": 1168
},
{
"epoch": 2.7776456599286563,
"grad_norm": 1.2622320191725491,
"learning_rate": 4.420429124268433e-06,
"loss": 1.78,
"step": 1169
},
{
"epoch": 2.780023781212842,
"grad_norm": 1.2724707119064946,
"learning_rate": 4.40454234150936e-06,
"loss": 1.6128,
"step": 1170
},
{
"epoch": 2.7824019024970275,
"grad_norm": 1.2162763695266292,
"learning_rate": 4.388676091519717e-06,
"loss": 1.7501,
"step": 1171
},
{
"epoch": 2.784780023781213,
"grad_norm": 1.0596805705972367,
"learning_rate": 4.372830432521377e-06,
"loss": 1.509,
"step": 1172
},
{
"epoch": 2.7871581450653986,
"grad_norm": 1.1796451830514862,
"learning_rate": 4.357005422660654e-06,
"loss": 1.6277,
"step": 1173
},
{
"epoch": 2.7895362663495837,
"grad_norm": 1.1869248634096565,
"learning_rate": 4.341201120008101e-06,
"loss": 1.5832,
"step": 1174
},
{
"epoch": 2.7919143876337693,
"grad_norm": 1.1768008704846398,
"learning_rate": 4.32541758255827e-06,
"loss": 1.5837,
"step": 1175
},
{
"epoch": 2.794292508917955,
"grad_norm": 1.359133421039981,
"learning_rate": 4.3096548682295304e-06,
"loss": 1.7285,
"step": 1176
},
{
"epoch": 2.7966706302021405,
"grad_norm": 1.1736998914426913,
"learning_rate": 4.293913034863827e-06,
"loss": 1.678,
"step": 1177
},
{
"epoch": 2.7990487514863256,
"grad_norm": 1.4230563916875643,
"learning_rate": 4.278192140226481e-06,
"loss": 1.8562,
"step": 1178
},
{
"epoch": 2.801426872770511,
"grad_norm": 1.3115643087054751,
"learning_rate": 4.262492242005993e-06,
"loss": 1.7907,
"step": 1179
},
{
"epoch": 2.8038049940546967,
"grad_norm": 1.2595762549552387,
"learning_rate": 4.2468133978137945e-06,
"loss": 1.7207,
"step": 1180
},
{
"epoch": 2.8061831153388823,
"grad_norm": 1.1249651556654061,
"learning_rate": 4.23115566518408e-06,
"loss": 1.7445,
"step": 1181
},
{
"epoch": 2.808561236623068,
"grad_norm": 5.900623578238023,
"learning_rate": 4.215519101573561e-06,
"loss": 1.6315,
"step": 1182
},
{
"epoch": 2.8109393579072535,
"grad_norm": 0.9990899222831929,
"learning_rate": 4.199903764361266e-06,
"loss": 1.7415,
"step": 1183
},
{
"epoch": 2.8133174791914386,
"grad_norm": 1.1614724591043748,
"learning_rate": 4.1843097108483465e-06,
"loss": 1.6034,
"step": 1184
},
{
"epoch": 2.815695600475624,
"grad_norm": 1.2526358529499493,
"learning_rate": 4.168736998257835e-06,
"loss": 1.618,
"step": 1185
},
{
"epoch": 2.8180737217598097,
"grad_norm": 1.1343545114240972,
"learning_rate": 4.153185683734471e-06,
"loss": 1.5081,
"step": 1186
},
{
"epoch": 2.8204518430439953,
"grad_norm": 1.236589459861005,
"learning_rate": 4.137655824344456e-06,
"loss": 1.7708,
"step": 1187
},
{
"epoch": 2.822829964328181,
"grad_norm": 1.1912640761030586,
"learning_rate": 4.12214747707527e-06,
"loss": 1.6344,
"step": 1188
},
{
"epoch": 2.825208085612366,
"grad_norm": 1.2734504646174627,
"learning_rate": 4.106660698835454e-06,
"loss": 1.7098,
"step": 1189
},
{
"epoch": 2.8275862068965516,
"grad_norm": 1.4358201608935164,
"learning_rate": 4.091195546454398e-06,
"loss": 1.8293,
"step": 1190
},
{
"epoch": 2.829964328180737,
"grad_norm": 1.2695044259379555,
"learning_rate": 4.07575207668213e-06,
"loss": 1.7176,
"step": 1191
},
{
"epoch": 2.8323424494649228,
"grad_norm": 1.1791999494249772,
"learning_rate": 4.060330346189125e-06,
"loss": 1.7079,
"step": 1192
},
{
"epoch": 2.8347205707491083,
"grad_norm": 1.0485457720571727,
"learning_rate": 4.044930411566072e-06,
"loss": 1.5055,
"step": 1193
},
{
"epoch": 2.837098692033294,
"grad_norm": 1.1504512570150962,
"learning_rate": 4.029552329323689e-06,
"loss": 1.7233,
"step": 1194
},
{
"epoch": 2.839476813317479,
"grad_norm": 1.1903698707292614,
"learning_rate": 4.014196155892503e-06,
"loss": 1.6478,
"step": 1195
},
{
"epoch": 2.8418549346016646,
"grad_norm": 1.1384007435698376,
"learning_rate": 3.998861947622635e-06,
"loss": 1.6235,
"step": 1196
},
{
"epoch": 2.84423305588585,
"grad_norm": 1.2028078323116806,
"learning_rate": 3.983549760783625e-06,
"loss": 1.6793,
"step": 1197
},
{
"epoch": 2.8466111771700358,
"grad_norm": 1.1401231552934252,
"learning_rate": 3.968259651564183e-06,
"loss": 1.5737,
"step": 1198
},
{
"epoch": 2.8489892984542213,
"grad_norm": 1.4317094122590954,
"learning_rate": 3.952991676072022e-06,
"loss": 1.7094,
"step": 1199
},
{
"epoch": 2.8513674197384065,
"grad_norm": 1.1867618341971813,
"learning_rate": 3.937745890333623e-06,
"loss": 1.6757,
"step": 1200
},
{
"epoch": 2.853745541022592,
"grad_norm": 1.24922668225119,
"learning_rate": 3.92252235029404e-06,
"loss": 1.7348,
"step": 1201
},
{
"epoch": 2.8561236623067776,
"grad_norm": 2.0482492936414047,
"learning_rate": 3.9073211118167065e-06,
"loss": 1.6081,
"step": 1202
},
{
"epoch": 2.858501783590963,
"grad_norm": 1.3060014185372235,
"learning_rate": 3.8921422306832046e-06,
"loss": 1.8234,
"step": 1203
},
{
"epoch": 2.8608799048751488,
"grad_norm": 1.0794119105154323,
"learning_rate": 3.876985762593091e-06,
"loss": 1.6398,
"step": 1204
},
{
"epoch": 2.8632580261593343,
"grad_norm": 1.0780158189794518,
"learning_rate": 3.861851763163665e-06,
"loss": 1.7754,
"step": 1205
},
{
"epoch": 2.8656361474435195,
"grad_norm": 1.4861159808624935,
"learning_rate": 3.846740287929778e-06,
"loss": 1.5971,
"step": 1206
},
{
"epoch": 2.868014268727705,
"grad_norm": 1.2055960321740489,
"learning_rate": 3.831651392343634e-06,
"loss": 1.8145,
"step": 1207
},
{
"epoch": 2.8703923900118906,
"grad_norm": 1.3800077509245512,
"learning_rate": 3.8165851317745705e-06,
"loss": 1.9161,
"step": 1208
},
{
"epoch": 2.872770511296076,
"grad_norm": 1.33689102839319,
"learning_rate": 3.801541561508878e-06,
"loss": 1.7726,
"step": 1209
},
{
"epoch": 2.8751486325802613,
"grad_norm": 1.2125563028361928,
"learning_rate": 3.7865207367495716e-06,
"loss": 1.82,
"step": 1210
},
{
"epoch": 2.877526753864447,
"grad_norm": 1.1442793595703613,
"learning_rate": 3.7715227126162025e-06,
"loss": 1.8318,
"step": 1211
},
{
"epoch": 2.8799048751486325,
"grad_norm": 1.1633615922891616,
"learning_rate": 3.756547544144664e-06,
"loss": 1.8416,
"step": 1212
},
{
"epoch": 2.882282996432818,
"grad_norm": 1.1373224437983922,
"learning_rate": 3.7415952862869654e-06,
"loss": 1.6247,
"step": 1213
},
{
"epoch": 2.8846611177170036,
"grad_norm": 1.137322810651517,
"learning_rate": 3.7266659939110585e-06,
"loss": 1.5931,
"step": 1214
},
{
"epoch": 2.887039239001189,
"grad_norm": 1.0775318054436687,
"learning_rate": 3.7117597218006105e-06,
"loss": 1.6027,
"step": 1215
},
{
"epoch": 2.8894173602853748,
"grad_norm": 1.3112661099152905,
"learning_rate": 3.69687652465482e-06,
"loss": 1.6443,
"step": 1216
},
{
"epoch": 2.89179548156956,
"grad_norm": 1.3918606030814118,
"learning_rate": 3.6820164570882146e-06,
"loss": 1.6602,
"step": 1217
},
{
"epoch": 2.8941736028537455,
"grad_norm": 1.3938215922035622,
"learning_rate": 3.667179573630437e-06,
"loss": 1.6631,
"step": 1218
},
{
"epoch": 2.896551724137931,
"grad_norm": 1.2301421785274869,
"learning_rate": 3.6523659287260695e-06,
"loss": 1.7729,
"step": 1219
},
{
"epoch": 2.8989298454221166,
"grad_norm": 3.1490088690528877,
"learning_rate": 3.6375755767344047e-06,
"loss": 1.614,
"step": 1220
},
{
"epoch": 2.9013079667063018,
"grad_norm": 1.3157024963066564,
"learning_rate": 3.6228085719292647e-06,
"loss": 1.4488,
"step": 1221
},
{
"epoch": 2.9036860879904873,
"grad_norm": 1.4102599859634863,
"learning_rate": 3.6080649684988077e-06,
"loss": 1.75,
"step": 1222
},
{
"epoch": 2.906064209274673,
"grad_norm": 1.14892092907497,
"learning_rate": 3.5933448205453104e-06,
"loss": 1.7228,
"step": 1223
},
{
"epoch": 2.9084423305588585,
"grad_norm": 1.136747625427715,
"learning_rate": 3.578648182084975e-06,
"loss": 1.4841,
"step": 1224
},
{
"epoch": 2.910820451843044,
"grad_norm": 1.138804446403443,
"learning_rate": 3.5639751070477478e-06,
"loss": 1.4907,
"step": 1225
},
{
"epoch": 2.9131985731272296,
"grad_norm": 1.2075949994709174,
"learning_rate": 3.5493256492770935e-06,
"loss": 1.8892,
"step": 1226
},
{
"epoch": 2.915576694411415,
"grad_norm": 1.3415354960606323,
"learning_rate": 3.5346998625298267e-06,
"loss": 1.8931,
"step": 1227
},
{
"epoch": 2.9179548156956003,
"grad_norm": 1.087555330181847,
"learning_rate": 3.520097800475889e-06,
"loss": 1.6262,
"step": 1228
},
{
"epoch": 2.920332936979786,
"grad_norm": 1.2882921866824912,
"learning_rate": 3.505519516698165e-06,
"loss": 1.8168,
"step": 1229
},
{
"epoch": 2.9227110582639715,
"grad_norm": 1.5314068926814943,
"learning_rate": 3.49096506469229e-06,
"loss": 1.8276,
"step": 1230
},
{
"epoch": 2.925089179548157,
"grad_norm": 1.2493103784228665,
"learning_rate": 3.476434497866439e-06,
"loss": 1.758,
"step": 1231
},
{
"epoch": 2.927467300832342,
"grad_norm": 1.3521688478182825,
"learning_rate": 3.4619278695411495e-06,
"loss": 1.7994,
"step": 1232
},
{
"epoch": 2.9298454221165278,
"grad_norm": 1.257370521239708,
"learning_rate": 3.447445232949107e-06,
"loss": 1.9069,
"step": 1233
},
{
"epoch": 2.9322235434007133,
"grad_norm": 1.5545256817738884,
"learning_rate": 3.4329866412349578e-06,
"loss": 1.6923,
"step": 1234
},
{
"epoch": 2.934601664684899,
"grad_norm": 1.357020107953056,
"learning_rate": 3.418552147455125e-06,
"loss": 1.8293,
"step": 1235
},
{
"epoch": 2.9369797859690845,
"grad_norm": 1.2639785875985403,
"learning_rate": 3.40414180457759e-06,
"loss": 1.6281,
"step": 1236
},
{
"epoch": 2.93935790725327,
"grad_norm": 1.0769659643827696,
"learning_rate": 3.389755665481723e-06,
"loss": 1.6815,
"step": 1237
},
{
"epoch": 2.9417360285374556,
"grad_norm": 1.1618459096532834,
"learning_rate": 3.3753937829580697e-06,
"loss": 1.6085,
"step": 1238
},
{
"epoch": 2.9441141498216408,
"grad_norm": 1.1040704664620735,
"learning_rate": 3.3610562097081646e-06,
"loss": 1.7131,
"step": 1239
},
{
"epoch": 2.9464922711058263,
"grad_norm": 1.2015097583626706,
"learning_rate": 3.3467429983443477e-06,
"loss": 1.6711,
"step": 1240
},
{
"epoch": 2.948870392390012,
"grad_norm": 1.2244144308823477,
"learning_rate": 3.33245420138955e-06,
"loss": 1.714,
"step": 1241
},
{
"epoch": 2.9512485136741975,
"grad_norm": 1.0863527927205638,
"learning_rate": 3.3181898712771242e-06,
"loss": 1.7266,
"step": 1242
},
{
"epoch": 2.9536266349583826,
"grad_norm": 1.1418737529822565,
"learning_rate": 3.3039500603506324e-06,
"loss": 1.7474,
"step": 1243
},
{
"epoch": 2.956004756242568,
"grad_norm": 1.5000461770427274,
"learning_rate": 3.2897348208636616e-06,
"loss": 1.8399,
"step": 1244
},
{
"epoch": 2.9583828775267538,
"grad_norm": 1.1758761049789366,
"learning_rate": 3.275544204979643e-06,
"loss": 1.6577,
"step": 1245
},
{
"epoch": 2.9607609988109393,
"grad_norm": 1.1225755305291578,
"learning_rate": 3.2613782647716374e-06,
"loss": 1.6723,
"step": 1246
},
{
"epoch": 2.963139120095125,
"grad_norm": 1.1070527347815056,
"learning_rate": 3.247237052222172e-06,
"loss": 1.4031,
"step": 1247
},
{
"epoch": 2.9655172413793105,
"grad_norm": 1.35321719646625,
"learning_rate": 3.233120619223021e-06,
"loss": 1.8406,
"step": 1248
},
{
"epoch": 2.967895362663496,
"grad_norm": 1.2364165288535383,
"learning_rate": 3.219029017575034e-06,
"loss": 1.6799,
"step": 1249
},
{
"epoch": 2.970273483947681,
"grad_norm": 1.2121445966430984,
"learning_rate": 3.2049622989879446e-06,
"loss": 1.8344,
"step": 1250
},
{
"epoch": 2.972651605231867,
"grad_norm": 1.2037195506097635,
"learning_rate": 3.19092051508017e-06,
"loss": 1.6278,
"step": 1251
},
{
"epoch": 2.9750297265160524,
"grad_norm": 1.2900489275647031,
"learning_rate": 3.1769037173786376e-06,
"loss": 1.6574,
"step": 1252
},
{
"epoch": 2.977407847800238,
"grad_norm": 1.3972721035539168,
"learning_rate": 3.1629119573185796e-06,
"loss": 1.8404,
"step": 1253
},
{
"epoch": 2.979785969084423,
"grad_norm": 1.332706172864179,
"learning_rate": 3.148945286243349e-06,
"loss": 1.6114,
"step": 1254
},
{
"epoch": 2.9821640903686086,
"grad_norm": 1.1696702073439071,
"learning_rate": 3.1350037554042446e-06,
"loss": 1.6369,
"step": 1255
},
{
"epoch": 2.984542211652794,
"grad_norm": 1.166223919795344,
"learning_rate": 3.1210874159603044e-06,
"loss": 1.3043,
"step": 1256
},
{
"epoch": 2.98692033293698,
"grad_norm": 1.2965140677555993,
"learning_rate": 3.1071963189781207e-06,
"loss": 1.6867,
"step": 1257
},
{
"epoch": 2.9892984542211654,
"grad_norm": 1.2624465537922454,
"learning_rate": 3.093330515431673e-06,
"loss": 1.8266,
"step": 1258
},
{
"epoch": 2.991676575505351,
"grad_norm": 1.4512366062742095,
"learning_rate": 3.079490056202107e-06,
"loss": 1.7582,
"step": 1259
},
{
"epoch": 2.9940546967895365,
"grad_norm": 1.2114616822405386,
"learning_rate": 3.065674992077584e-06,
"loss": 1.4938,
"step": 1260
},
{
"epoch": 2.9940546967895365,
"eval_loss": 2.375544786453247,
"eval_runtime": 65.6032,
"eval_samples_per_second": 9.74,
"eval_steps_per_second": 1.219,
"step": 1260
}
],
"logging_steps": 1,
"max_steps": 1680,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 210,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 7.810082245674271e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}