Hasnonname's picture
Add files using upload-large-folder tool
080cd68 verified
raw
history blame
99.9 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9979879275653922,
"eval_steps": 47,
"global_step": 558,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0053655264922870555,
"grad_norm": 1.2779298491774165,
"learning_rate": 1.8518518518518519e-06,
"loss": 2.0226,
"step": 1
},
{
"epoch": 0.0053655264922870555,
"eval_loss": 3.5763909816741943,
"eval_runtime": 403.4915,
"eval_samples_per_second": 6.317,
"eval_steps_per_second": 0.791,
"step": 1
},
{
"epoch": 0.010731052984574111,
"grad_norm": 0.9868395151451838,
"learning_rate": 3.7037037037037037e-06,
"loss": 2.0251,
"step": 2
},
{
"epoch": 0.01609657947686117,
"grad_norm": 1.124719320101948,
"learning_rate": 5.555555555555556e-06,
"loss": 2.0676,
"step": 3
},
{
"epoch": 0.021462105969148222,
"grad_norm": 0.8967993867216901,
"learning_rate": 7.4074074074074075e-06,
"loss": 2.0685,
"step": 4
},
{
"epoch": 0.02682763246143528,
"grad_norm": 0.8530329942799596,
"learning_rate": 9.259259259259259e-06,
"loss": 1.975,
"step": 5
},
{
"epoch": 0.03219315895372234,
"grad_norm": 17.572932401470002,
"learning_rate": 1.1111111111111112e-05,
"loss": 2.0319,
"step": 6
},
{
"epoch": 0.03755868544600939,
"grad_norm": 1.5955614660997672,
"learning_rate": 1.2962962962962962e-05,
"loss": 1.9776,
"step": 7
},
{
"epoch": 0.042924211938296444,
"grad_norm": 1.0305843114650708,
"learning_rate": 1.4814814814814815e-05,
"loss": 1.9384,
"step": 8
},
{
"epoch": 0.0482897384305835,
"grad_norm": 0.7931768664816836,
"learning_rate": 1.6666666666666667e-05,
"loss": 2.0357,
"step": 9
},
{
"epoch": 0.05365526492287056,
"grad_norm": 0.7458181490503686,
"learning_rate": 1.8518518518518518e-05,
"loss": 1.9343,
"step": 10
},
{
"epoch": 0.05902079141515761,
"grad_norm": 1.0379874858551188,
"learning_rate": 2.037037037037037e-05,
"loss": 1.9069,
"step": 11
},
{
"epoch": 0.06438631790744467,
"grad_norm": 0.9634098196921086,
"learning_rate": 2.2222222222222223e-05,
"loss": 1.8831,
"step": 12
},
{
"epoch": 0.06975184439973173,
"grad_norm": 0.8492217551097285,
"learning_rate": 2.4074074074074074e-05,
"loss": 1.9665,
"step": 13
},
{
"epoch": 0.07511737089201878,
"grad_norm": 0.8866414055564286,
"learning_rate": 2.5925925925925925e-05,
"loss": 1.9859,
"step": 14
},
{
"epoch": 0.08048289738430583,
"grad_norm": 0.8975616587977772,
"learning_rate": 2.777777777777778e-05,
"loss": 1.9621,
"step": 15
},
{
"epoch": 0.08584842387659289,
"grad_norm": 0.8856060845121474,
"learning_rate": 2.962962962962963e-05,
"loss": 2.0504,
"step": 16
},
{
"epoch": 0.09121395036887994,
"grad_norm": 0.7934430921825818,
"learning_rate": 3.148148148148148e-05,
"loss": 1.8965,
"step": 17
},
{
"epoch": 0.096579476861167,
"grad_norm": 0.8411958882279533,
"learning_rate": 3.3333333333333335e-05,
"loss": 1.834,
"step": 18
},
{
"epoch": 0.10194500335345406,
"grad_norm": 0.9605732267993595,
"learning_rate": 3.518518518518519e-05,
"loss": 2.0321,
"step": 19
},
{
"epoch": 0.10731052984574112,
"grad_norm": 0.8964432934098547,
"learning_rate": 3.7037037037037037e-05,
"loss": 1.9165,
"step": 20
},
{
"epoch": 0.11267605633802817,
"grad_norm": 0.8325547065304583,
"learning_rate": 3.888888888888889e-05,
"loss": 2.0088,
"step": 21
},
{
"epoch": 0.11804158283031523,
"grad_norm": 0.7337982135739225,
"learning_rate": 4.074074074074074e-05,
"loss": 1.8652,
"step": 22
},
{
"epoch": 0.12340710932260228,
"grad_norm": 0.8703749413838572,
"learning_rate": 4.259259259259259e-05,
"loss": 1.9384,
"step": 23
},
{
"epoch": 0.12877263581488935,
"grad_norm": 0.752681510079145,
"learning_rate": 4.4444444444444447e-05,
"loss": 1.84,
"step": 24
},
{
"epoch": 0.1341381623071764,
"grad_norm": 0.7664305287185245,
"learning_rate": 4.62962962962963e-05,
"loss": 1.917,
"step": 25
},
{
"epoch": 0.13950368879946345,
"grad_norm": 0.7376417365374465,
"learning_rate": 4.814814814814815e-05,
"loss": 1.8995,
"step": 26
},
{
"epoch": 0.1448692152917505,
"grad_norm": 0.6723119101020634,
"learning_rate": 5e-05,
"loss": 1.8572,
"step": 27
},
{
"epoch": 0.15023474178403756,
"grad_norm": 0.6636335261172684,
"learning_rate": 4.999956245830044e-05,
"loss": 1.8311,
"step": 28
},
{
"epoch": 0.15560026827632462,
"grad_norm": 0.7118137563548644,
"learning_rate": 4.9998249848517185e-05,
"loss": 1.9461,
"step": 29
},
{
"epoch": 0.16096579476861167,
"grad_norm": 0.5571924001587539,
"learning_rate": 4.999606221659595e-05,
"loss": 1.9125,
"step": 30
},
{
"epoch": 0.16633132126089872,
"grad_norm": 0.5895748551871794,
"learning_rate": 4.999299963911115e-05,
"loss": 1.9295,
"step": 31
},
{
"epoch": 0.17169684775318578,
"grad_norm": 0.5114717351226575,
"learning_rate": 4.9989062223263216e-05,
"loss": 1.8642,
"step": 32
},
{
"epoch": 0.17706237424547283,
"grad_norm": 0.5658751591114588,
"learning_rate": 4.998425010687484e-05,
"loss": 1.9465,
"step": 33
},
{
"epoch": 0.18242790073775988,
"grad_norm": 0.4198888613535172,
"learning_rate": 4.997856345838615e-05,
"loss": 1.7824,
"step": 34
},
{
"epoch": 0.18779342723004694,
"grad_norm": 0.558439585084249,
"learning_rate": 4.99720024768488e-05,
"loss": 1.9157,
"step": 35
},
{
"epoch": 0.193158953722334,
"grad_norm": 0.4888963940737741,
"learning_rate": 4.996456739191905e-05,
"loss": 1.9136,
"step": 36
},
{
"epoch": 0.19852448021462105,
"grad_norm": 0.5106753170932593,
"learning_rate": 4.995625846384966e-05,
"loss": 1.8623,
"step": 37
},
{
"epoch": 0.20389000670690813,
"grad_norm": 0.5942581620888907,
"learning_rate": 4.994707598348085e-05,
"loss": 1.8741,
"step": 38
},
{
"epoch": 0.20925553319919518,
"grad_norm": 0.5106367767120957,
"learning_rate": 4.993702027223004e-05,
"loss": 1.9413,
"step": 39
},
{
"epoch": 0.21462105969148224,
"grad_norm": 0.5113103948113709,
"learning_rate": 4.992609168208069e-05,
"loss": 1.9966,
"step": 40
},
{
"epoch": 0.2199865861837693,
"grad_norm": 0.5212461839914251,
"learning_rate": 4.9914290595569895e-05,
"loss": 1.8474,
"step": 41
},
{
"epoch": 0.22535211267605634,
"grad_norm": 0.5161668483087154,
"learning_rate": 4.9901617425775067e-05,
"loss": 1.8557,
"step": 42
},
{
"epoch": 0.2307176391683434,
"grad_norm": 0.5484719777790631,
"learning_rate": 4.988807261629942e-05,
"loss": 1.8769,
"step": 43
},
{
"epoch": 0.23608316566063045,
"grad_norm": 0.4937411819105057,
"learning_rate": 4.987365664125647e-05,
"loss": 1.8923,
"step": 44
},
{
"epoch": 0.2414486921529175,
"grad_norm": 0.5804971756745295,
"learning_rate": 4.985837000525343e-05,
"loss": 1.954,
"step": 45
},
{
"epoch": 0.24681421864520456,
"grad_norm": 0.531027559174979,
"learning_rate": 4.984221324337356e-05,
"loss": 1.9758,
"step": 46
},
{
"epoch": 0.25217974513749164,
"grad_norm": 0.5132447531432143,
"learning_rate": 4.982518692115744e-05,
"loss": 1.8588,
"step": 47
},
{
"epoch": 0.25217974513749164,
"eval_loss": 2.2616188526153564,
"eval_runtime": 403.9965,
"eval_samples_per_second": 6.309,
"eval_steps_per_second": 0.79,
"step": 47
},
{
"epoch": 0.2575452716297787,
"grad_norm": 0.4799971353542736,
"learning_rate": 4.980729163458312e-05,
"loss": 1.8558,
"step": 48
},
{
"epoch": 0.26291079812206575,
"grad_norm": 0.4318157242619782,
"learning_rate": 4.978852801004534e-05,
"loss": 1.9574,
"step": 49
},
{
"epoch": 0.2682763246143528,
"grad_norm": 0.44416051699732234,
"learning_rate": 4.976889670433355e-05,
"loss": 1.7988,
"step": 50
},
{
"epoch": 0.27364185110663986,
"grad_norm": 0.42380891285410505,
"learning_rate": 4.974839840460895e-05,
"loss": 1.8296,
"step": 51
},
{
"epoch": 0.2790073775989269,
"grad_norm": 0.4989220387491634,
"learning_rate": 4.97270338283804e-05,
"loss": 1.8996,
"step": 52
},
{
"epoch": 0.28437290409121396,
"grad_norm": 0.44554710351703636,
"learning_rate": 4.970480372347934e-05,
"loss": 1.9553,
"step": 53
},
{
"epoch": 0.289738430583501,
"grad_norm": 0.5597504355268026,
"learning_rate": 4.9681708868033616e-05,
"loss": 1.9979,
"step": 54
},
{
"epoch": 0.29510395707578807,
"grad_norm": 0.4758356280331154,
"learning_rate": 4.9657750070440196e-05,
"loss": 1.9244,
"step": 55
},
{
"epoch": 0.3004694835680751,
"grad_norm": 0.4337532785073617,
"learning_rate": 4.963292816933692e-05,
"loss": 1.9099,
"step": 56
},
{
"epoch": 0.3058350100603622,
"grad_norm": 0.48193864337289033,
"learning_rate": 4.9607244033573156e-05,
"loss": 1.8076,
"step": 57
},
{
"epoch": 0.31120053655264923,
"grad_norm": 0.5180685237569209,
"learning_rate": 4.9580698562179297e-05,
"loss": 1.9031,
"step": 58
},
{
"epoch": 0.3165660630449363,
"grad_norm": 0.4227856708321325,
"learning_rate": 4.955329268433543e-05,
"loss": 1.8111,
"step": 59
},
{
"epoch": 0.32193158953722334,
"grad_norm": 0.4457771321490526,
"learning_rate": 4.9525027359338696e-05,
"loss": 1.8771,
"step": 60
},
{
"epoch": 0.3272971160295104,
"grad_norm": 0.383809821251567,
"learning_rate": 4.949590357656975e-05,
"loss": 1.8384,
"step": 61
},
{
"epoch": 0.33266264252179745,
"grad_norm": 0.5245332667935808,
"learning_rate": 4.946592235545815e-05,
"loss": 1.948,
"step": 62
},
{
"epoch": 0.3380281690140845,
"grad_norm": 0.48131587989680386,
"learning_rate": 4.9435084745446666e-05,
"loss": 1.9546,
"step": 63
},
{
"epoch": 0.34339369550637155,
"grad_norm": 0.47579745445020977,
"learning_rate": 4.940339182595451e-05,
"loss": 1.9085,
"step": 64
},
{
"epoch": 0.3487592219986586,
"grad_norm": 0.5162389548444078,
"learning_rate": 4.9370844706339594e-05,
"loss": 1.9263,
"step": 65
},
{
"epoch": 0.35412474849094566,
"grad_norm": 0.4430593262438425,
"learning_rate": 4.933744452585966e-05,
"loss": 1.8264,
"step": 66
},
{
"epoch": 0.3594902749832327,
"grad_norm": 0.4564426740705147,
"learning_rate": 4.930319245363248e-05,
"loss": 1.7974,
"step": 67
},
{
"epoch": 0.36485580147551977,
"grad_norm": 0.4896650401826788,
"learning_rate": 4.926808968859483e-05,
"loss": 1.8567,
"step": 68
},
{
"epoch": 0.3702213279678068,
"grad_norm": 0.4051071793789495,
"learning_rate": 4.923213745946059e-05,
"loss": 1.8201,
"step": 69
},
{
"epoch": 0.3755868544600939,
"grad_norm": 0.48978661474326063,
"learning_rate": 4.919533702467771e-05,
"loss": 1.8006,
"step": 70
},
{
"epoch": 0.38095238095238093,
"grad_norm": 0.5175966382748892,
"learning_rate": 4.9157689672384174e-05,
"loss": 1.7904,
"step": 71
},
{
"epoch": 0.386317907444668,
"grad_norm": 0.49081107125596507,
"learning_rate": 4.91191967203629e-05,
"loss": 1.9606,
"step": 72
},
{
"epoch": 0.39168343393695504,
"grad_norm": 0.5920859101699153,
"learning_rate": 4.907985951599563e-05,
"loss": 1.8124,
"step": 73
},
{
"epoch": 0.3970489604292421,
"grad_norm": 0.4106853545996278,
"learning_rate": 4.9039679436215734e-05,
"loss": 1.9402,
"step": 74
},
{
"epoch": 0.4024144869215292,
"grad_norm": 0.6429573337250454,
"learning_rate": 4.899865788746005e-05,
"loss": 1.8233,
"step": 75
},
{
"epoch": 0.40778001341381626,
"grad_norm": 0.5770635817407002,
"learning_rate": 4.895679630561963e-05,
"loss": 1.8909,
"step": 76
},
{
"epoch": 0.4131455399061033,
"grad_norm": 0.8202937145800111,
"learning_rate": 4.891409615598949e-05,
"loss": 1.8635,
"step": 77
},
{
"epoch": 0.41851106639839036,
"grad_norm": 0.6871160286583309,
"learning_rate": 4.88705589332173e-05,
"loss": 1.9497,
"step": 78
},
{
"epoch": 0.4238765928906774,
"grad_norm": 0.4438407032445468,
"learning_rate": 4.882618616125111e-05,
"loss": 1.9336,
"step": 79
},
{
"epoch": 0.42924211938296447,
"grad_norm": 0.545225076259868,
"learning_rate": 4.878097939328596e-05,
"loss": 1.843,
"step": 80
},
{
"epoch": 0.4346076458752515,
"grad_norm": 0.5120969176703908,
"learning_rate": 4.873494021170953e-05,
"loss": 1.9019,
"step": 81
},
{
"epoch": 0.4399731723675386,
"grad_norm": 0.41579446347477983,
"learning_rate": 4.868807022804678e-05,
"loss": 1.8617,
"step": 82
},
{
"epoch": 0.44533869885982563,
"grad_norm": 0.5510693357302907,
"learning_rate": 4.864037108290347e-05,
"loss": 1.9433,
"step": 83
},
{
"epoch": 0.4507042253521127,
"grad_norm": 0.40913367019675556,
"learning_rate": 4.859184444590882e-05,
"loss": 1.8353,
"step": 84
},
{
"epoch": 0.45606975184439974,
"grad_norm": 0.3763529830498608,
"learning_rate": 4.854249201565701e-05,
"loss": 1.7299,
"step": 85
},
{
"epoch": 0.4614352783366868,
"grad_norm": 0.47448012995411376,
"learning_rate": 4.849231551964771e-05,
"loss": 1.8547,
"step": 86
},
{
"epoch": 0.46680080482897385,
"grad_norm": 0.4181797287844973,
"learning_rate": 4.84413167142257e-05,
"loss": 1.8186,
"step": 87
},
{
"epoch": 0.4721663313212609,
"grad_norm": 0.4781785535505977,
"learning_rate": 4.838949738451929e-05,
"loss": 1.8965,
"step": 88
},
{
"epoch": 0.47753185781354796,
"grad_norm": 0.4351347675471463,
"learning_rate": 4.833685934437787e-05,
"loss": 1.863,
"step": 89
},
{
"epoch": 0.482897384305835,
"grad_norm": 0.3874884362959953,
"learning_rate": 4.8283404436308464e-05,
"loss": 1.7396,
"step": 90
},
{
"epoch": 0.48826291079812206,
"grad_norm": 0.49444219921157534,
"learning_rate": 4.8229134531411166e-05,
"loss": 1.8444,
"step": 91
},
{
"epoch": 0.4936284372904091,
"grad_norm": 0.39922493357750344,
"learning_rate": 4.8174051529313704e-05,
"loss": 1.8247,
"step": 92
},
{
"epoch": 0.49899396378269617,
"grad_norm": 0.5511215543224721,
"learning_rate": 4.81181573581049e-05,
"loss": 1.9533,
"step": 93
},
{
"epoch": 0.5043594902749833,
"grad_norm": 0.41108793496635143,
"learning_rate": 4.8061453974267195e-05,
"loss": 1.7952,
"step": 94
},
{
"epoch": 0.5043594902749833,
"eval_loss": 2.198935031890869,
"eval_runtime": 403.21,
"eval_samples_per_second": 6.322,
"eval_steps_per_second": 0.791,
"step": 94
},
{
"epoch": 0.5097250167672703,
"grad_norm": 0.43234228757508586,
"learning_rate": 4.80039433626082e-05,
"loss": 1.8667,
"step": 95
},
{
"epoch": 0.5150905432595574,
"grad_norm": 0.40707489762011484,
"learning_rate": 4.7945627536191166e-05,
"loss": 1.9081,
"step": 96
},
{
"epoch": 0.5204560697518444,
"grad_norm": 0.41662703960667113,
"learning_rate": 4.788650853626456e-05,
"loss": 1.7536,
"step": 97
},
{
"epoch": 0.5258215962441315,
"grad_norm": 0.39754454922458654,
"learning_rate": 4.7826588432190614e-05,
"loss": 1.8851,
"step": 98
},
{
"epoch": 0.5311871227364185,
"grad_norm": 0.41218018944985396,
"learning_rate": 4.7765869321372836e-05,
"loss": 1.8062,
"step": 99
},
{
"epoch": 0.5365526492287056,
"grad_norm": 0.41546782428712065,
"learning_rate": 4.7704353329182673e-05,
"loss": 1.8491,
"step": 100
},
{
"epoch": 0.5419181757209927,
"grad_norm": 0.47175537335479406,
"learning_rate": 4.7642042608885064e-05,
"loss": 1.7917,
"step": 101
},
{
"epoch": 0.5472837022132797,
"grad_norm": 0.4373071645022652,
"learning_rate": 4.7578939341563095e-05,
"loss": 1.8632,
"step": 102
},
{
"epoch": 0.5526492287055668,
"grad_norm": 0.4369286574512984,
"learning_rate": 4.751504573604162e-05,
"loss": 1.8377,
"step": 103
},
{
"epoch": 0.5580147551978538,
"grad_norm": 0.40840919093725575,
"learning_rate": 4.745036402880999e-05,
"loss": 1.8397,
"step": 104
},
{
"epoch": 0.5633802816901409,
"grad_norm": 0.4023593789216688,
"learning_rate": 4.738489648394373e-05,
"loss": 1.9389,
"step": 105
},
{
"epoch": 0.5687458081824279,
"grad_norm": 0.45139851873493203,
"learning_rate": 4.731864539302531e-05,
"loss": 1.9384,
"step": 106
},
{
"epoch": 0.574111334674715,
"grad_norm": 0.518973698127367,
"learning_rate": 4.725161307506391e-05,
"loss": 1.8388,
"step": 107
},
{
"epoch": 0.579476861167002,
"grad_norm": 0.41440109186821383,
"learning_rate": 4.7183801876414294e-05,
"loss": 1.9807,
"step": 108
},
{
"epoch": 0.5848423876592891,
"grad_norm": 0.485066583270121,
"learning_rate": 4.711521417069462e-05,
"loss": 1.9518,
"step": 109
},
{
"epoch": 0.5902079141515761,
"grad_norm": 0.453283667200083,
"learning_rate": 4.70458523587034e-05,
"loss": 1.8874,
"step": 110
},
{
"epoch": 0.5955734406438632,
"grad_norm": 0.40638037542695366,
"learning_rate": 4.697571886833544e-05,
"loss": 1.8138,
"step": 111
},
{
"epoch": 0.6009389671361502,
"grad_norm": 0.47718260435002885,
"learning_rate": 4.6904816154496854e-05,
"loss": 1.8409,
"step": 112
},
{
"epoch": 0.6063044936284373,
"grad_norm": 0.45172842172725275,
"learning_rate": 4.683314669901918e-05,
"loss": 1.9261,
"step": 113
},
{
"epoch": 0.6116700201207244,
"grad_norm": 0.4713867113135493,
"learning_rate": 4.676071301057243e-05,
"loss": 1.8502,
"step": 114
},
{
"epoch": 0.6170355466130114,
"grad_norm": 0.42451831321750794,
"learning_rate": 4.668751762457734e-05,
"loss": 1.8489,
"step": 115
},
{
"epoch": 0.6224010731052985,
"grad_norm": 0.3958564758105134,
"learning_rate": 4.6613563103116594e-05,
"loss": 1.8922,
"step": 116
},
{
"epoch": 0.6277665995975855,
"grad_norm": 0.4009292301802374,
"learning_rate": 4.653885203484515e-05,
"loss": 1.8685,
"step": 117
},
{
"epoch": 0.6331321260898726,
"grad_norm": 0.37598195515170946,
"learning_rate": 4.6463387034899645e-05,
"loss": 1.7889,
"step": 118
},
{
"epoch": 0.6384976525821596,
"grad_norm": 0.4710597532282601,
"learning_rate": 4.638717074480682e-05,
"loss": 1.8739,
"step": 119
},
{
"epoch": 0.6438631790744467,
"grad_norm": 0.4146663267990981,
"learning_rate": 4.631020583239107e-05,
"loss": 1.8988,
"step": 120
},
{
"epoch": 0.6492287055667337,
"grad_norm": 0.4354294422619394,
"learning_rate": 4.6232494991681094e-05,
"loss": 1.8517,
"step": 121
},
{
"epoch": 0.6545942320590208,
"grad_norm": 0.5470562094587934,
"learning_rate": 4.615404094281554e-05,
"loss": 1.812,
"step": 122
},
{
"epoch": 0.6599597585513078,
"grad_norm": 0.37147901827060426,
"learning_rate": 4.607484643194788e-05,
"loss": 1.8447,
"step": 123
},
{
"epoch": 0.6653252850435949,
"grad_norm": 0.462225949803499,
"learning_rate": 4.599491423115014e-05,
"loss": 1.8442,
"step": 124
},
{
"epoch": 0.670690811535882,
"grad_norm": 0.4600536423365438,
"learning_rate": 4.5914247138316025e-05,
"loss": 1.926,
"step": 125
},
{
"epoch": 0.676056338028169,
"grad_norm": 0.4049920788749757,
"learning_rate": 4.5832847977062874e-05,
"loss": 1.8001,
"step": 126
},
{
"epoch": 0.681421864520456,
"grad_norm": 0.4052262687972146,
"learning_rate": 4.5750719596632885e-05,
"loss": 1.8616,
"step": 127
},
{
"epoch": 0.6867873910127431,
"grad_norm": 0.455611587260568,
"learning_rate": 4.5667864871793345e-05,
"loss": 1.7792,
"step": 128
},
{
"epoch": 0.6921529175050302,
"grad_norm": 0.4321930788528261,
"learning_rate": 4.558428670273601e-05,
"loss": 1.911,
"step": 129
},
{
"epoch": 0.6975184439973172,
"grad_norm": 0.3922798652619637,
"learning_rate": 4.549998801497564e-05,
"loss": 1.8979,
"step": 130
},
{
"epoch": 0.7028839704896043,
"grad_norm": 0.4338149926251678,
"learning_rate": 4.54149717592475e-05,
"loss": 1.793,
"step": 131
},
{
"epoch": 0.7082494969818913,
"grad_norm": 0.42943405977885124,
"learning_rate": 4.532924091140417e-05,
"loss": 1.8606,
"step": 132
},
{
"epoch": 0.7136150234741784,
"grad_norm": 0.4145321079739444,
"learning_rate": 4.524279847231131e-05,
"loss": 1.906,
"step": 133
},
{
"epoch": 0.7189805499664654,
"grad_norm": 0.38959038753607894,
"learning_rate": 4.515564746774265e-05,
"loss": 1.7323,
"step": 134
},
{
"epoch": 0.7243460764587525,
"grad_norm": 0.49146581928185773,
"learning_rate": 4.5067790948274094e-05,
"loss": 1.8303,
"step": 135
},
{
"epoch": 0.7297116029510395,
"grad_norm": 0.4638570517033698,
"learning_rate": 4.49792319891769e-05,
"loss": 1.8243,
"step": 136
},
{
"epoch": 0.7350771294433266,
"grad_norm": 0.38160033679573113,
"learning_rate": 4.4889973690310085e-05,
"loss": 1.8186,
"step": 137
},
{
"epoch": 0.7404426559356136,
"grad_norm": 0.5253837731077495,
"learning_rate": 4.480001917601185e-05,
"loss": 1.9086,
"step": 138
},
{
"epoch": 0.7458081824279007,
"grad_norm": 0.5010255463058967,
"learning_rate": 4.470937159499029e-05,
"loss": 1.9014,
"step": 139
},
{
"epoch": 0.7511737089201878,
"grad_norm": 0.44408252888412136,
"learning_rate": 4.461803412021314e-05,
"loss": 1.886,
"step": 140
},
{
"epoch": 0.7565392354124748,
"grad_norm": 0.5189286834612553,
"learning_rate": 4.4526009948796703e-05,
"loss": 1.8266,
"step": 141
},
{
"epoch": 0.7565392354124748,
"eval_loss": 2.1618716716766357,
"eval_runtime": 403.938,
"eval_samples_per_second": 6.31,
"eval_steps_per_second": 0.79,
"step": 141
},
{
"epoch": 0.7619047619047619,
"grad_norm": 0.4334524920814895,
"learning_rate": 4.4433302301893987e-05,
"loss": 1.8863,
"step": 142
},
{
"epoch": 0.7672702883970489,
"grad_norm": 0.4809144135248221,
"learning_rate": 4.433991442458188e-05,
"loss": 1.9075,
"step": 143
},
{
"epoch": 0.772635814889336,
"grad_norm": 0.5258910256081301,
"learning_rate": 4.4245849585747654e-05,
"loss": 1.9616,
"step": 144
},
{
"epoch": 0.778001341381623,
"grad_norm": 0.5131536872616902,
"learning_rate": 4.415111107797445e-05,
"loss": 1.8953,
"step": 145
},
{
"epoch": 0.7833668678739101,
"grad_norm": 0.45216150555891305,
"learning_rate": 4.4055702217426084e-05,
"loss": 1.8867,
"step": 146
},
{
"epoch": 0.7887323943661971,
"grad_norm": 0.47907057676705184,
"learning_rate": 4.395962634373097e-05,
"loss": 1.8335,
"step": 147
},
{
"epoch": 0.7940979208584842,
"grad_norm": 0.4945502094720473,
"learning_rate": 4.386288681986516e-05,
"loss": 1.9218,
"step": 148
},
{
"epoch": 0.7994634473507712,
"grad_norm": 0.4595838282230734,
"learning_rate": 4.376548703203474e-05,
"loss": 1.8603,
"step": 149
},
{
"epoch": 0.8048289738430584,
"grad_norm": 0.45448249598162854,
"learning_rate": 4.36674303895572e-05,
"loss": 1.8615,
"step": 150
},
{
"epoch": 0.8101945003353455,
"grad_norm": 0.4238860646236377,
"learning_rate": 4.356872032474213e-05,
"loss": 1.8374,
"step": 151
},
{
"epoch": 0.8155600268276325,
"grad_norm": 0.4519762198634726,
"learning_rate": 4.34693602927711e-05,
"loss": 1.8174,
"step": 152
},
{
"epoch": 0.8209255533199196,
"grad_norm": 0.45624612554196736,
"learning_rate": 4.336935377157668e-05,
"loss": 1.9105,
"step": 153
},
{
"epoch": 0.8262910798122066,
"grad_norm": 0.4076679531263349,
"learning_rate": 4.326870426172075e-05,
"loss": 1.8865,
"step": 154
},
{
"epoch": 0.8316566063044937,
"grad_norm": 0.4514412896189259,
"learning_rate": 4.3167415286271905e-05,
"loss": 1.9272,
"step": 155
},
{
"epoch": 0.8370221327967807,
"grad_norm": 0.40433175857719245,
"learning_rate": 4.3065490390682186e-05,
"loss": 1.814,
"step": 156
},
{
"epoch": 0.8423876592890678,
"grad_norm": 0.4159285646615437,
"learning_rate": 4.296293314266294e-05,
"loss": 1.8064,
"step": 157
},
{
"epoch": 0.8477531857813548,
"grad_norm": 0.4270220599999327,
"learning_rate": 4.2859747132060006e-05,
"loss": 1.7482,
"step": 158
},
{
"epoch": 0.8531187122736419,
"grad_norm": 0.5162993182758251,
"learning_rate": 4.275593597072796e-05,
"loss": 1.7763,
"step": 159
},
{
"epoch": 0.8584842387659289,
"grad_norm": 0.4065368416357515,
"learning_rate": 4.265150329240376e-05,
"loss": 1.885,
"step": 160
},
{
"epoch": 0.863849765258216,
"grad_norm": 0.5523980888304064,
"learning_rate": 4.2546452752579536e-05,
"loss": 1.971,
"step": 161
},
{
"epoch": 0.869215291750503,
"grad_norm": 0.4121779650999884,
"learning_rate": 4.2440788028374624e-05,
"loss": 1.9301,
"step": 162
},
{
"epoch": 0.8745808182427901,
"grad_norm": 0.4612759438023989,
"learning_rate": 4.233451281840686e-05,
"loss": 1.8564,
"step": 163
},
{
"epoch": 0.8799463447350772,
"grad_norm": 0.3977527412707747,
"learning_rate": 4.2227630842663136e-05,
"loss": 1.8876,
"step": 164
},
{
"epoch": 0.8853118712273642,
"grad_norm": 0.45276964005664955,
"learning_rate": 4.212014584236914e-05,
"loss": 1.8098,
"step": 165
},
{
"epoch": 0.8906773977196513,
"grad_norm": 0.37349476483277844,
"learning_rate": 4.2012061579858465e-05,
"loss": 1.8247,
"step": 166
},
{
"epoch": 0.8960429242119383,
"grad_norm": 0.39676844111499676,
"learning_rate": 4.190338183844086e-05,
"loss": 1.914,
"step": 167
},
{
"epoch": 0.9014084507042254,
"grad_norm": 0.3625691966854791,
"learning_rate": 4.1794110422269825e-05,
"loss": 1.92,
"step": 168
},
{
"epoch": 0.9067739771965124,
"grad_norm": 0.39688352111910685,
"learning_rate": 4.168425115620944e-05,
"loss": 1.8103,
"step": 169
},
{
"epoch": 0.9121395036887995,
"grad_norm": 0.3737222783168723,
"learning_rate": 4.157380788570053e-05,
"loss": 1.8215,
"step": 170
},
{
"epoch": 0.9175050301810865,
"grad_norm": 0.43842445138079555,
"learning_rate": 4.146278447662597e-05,
"loss": 1.8029,
"step": 171
},
{
"epoch": 0.9228705566733736,
"grad_norm": 0.47414715687042996,
"learning_rate": 4.1351184815175456e-05,
"loss": 1.8974,
"step": 172
},
{
"epoch": 0.9282360831656606,
"grad_norm": 0.4146539941665616,
"learning_rate": 4.123901280770945e-05,
"loss": 1.8871,
"step": 173
},
{
"epoch": 0.9336016096579477,
"grad_norm": 0.4088656394439229,
"learning_rate": 4.112627238062239e-05,
"loss": 1.8594,
"step": 174
},
{
"epoch": 0.9389671361502347,
"grad_norm": 0.40239314445747604,
"learning_rate": 4.101296748020533e-05,
"loss": 1.8207,
"step": 175
},
{
"epoch": 0.9443326626425218,
"grad_norm": 0.36389736696587754,
"learning_rate": 4.089910207250778e-05,
"loss": 1.8423,
"step": 176
},
{
"epoch": 0.9496981891348089,
"grad_norm": 0.41161031740209597,
"learning_rate": 4.0784680143198836e-05,
"loss": 1.8954,
"step": 177
},
{
"epoch": 0.9550637156270959,
"grad_norm": 0.3613771597005119,
"learning_rate": 4.0669705697427754e-05,
"loss": 1.9036,
"step": 178
},
{
"epoch": 0.960429242119383,
"grad_norm": 0.36880689305175635,
"learning_rate": 4.055418275968368e-05,
"loss": 1.8676,
"step": 179
},
{
"epoch": 0.96579476861167,
"grad_norm": 0.3892009514188599,
"learning_rate": 4.04381153736548e-05,
"loss": 1.8545,
"step": 180
},
{
"epoch": 0.9711602951039571,
"grad_norm": 0.39386413560247385,
"learning_rate": 4.032150760208684e-05,
"loss": 1.9306,
"step": 181
},
{
"epoch": 0.9765258215962441,
"grad_norm": 0.38085985815329654,
"learning_rate": 4.02043635266408e-05,
"loss": 1.9339,
"step": 182
},
{
"epoch": 0.9818913480885312,
"grad_norm": 0.3502551322824309,
"learning_rate": 4.00866872477501e-05,
"loss": 1.8438,
"step": 183
},
{
"epoch": 0.9872568745808182,
"grad_norm": 0.42923155491093273,
"learning_rate": 3.9968482884477075e-05,
"loss": 1.8326,
"step": 184
},
{
"epoch": 0.9926224010731053,
"grad_norm": 0.37629854526288287,
"learning_rate": 3.9849754574368766e-05,
"loss": 1.8521,
"step": 185
},
{
"epoch": 0.9979879275653923,
"grad_norm": 0.43648587047407056,
"learning_rate": 3.973050647331209e-05,
"loss": 1.9192,
"step": 186
},
{
"epoch": 1.0053655264922872,
"grad_norm": 1.1037688215019685,
"learning_rate": 3.9610742755388406e-05,
"loss": 3.5179,
"step": 187
},
{
"epoch": 1.010731052984574,
"grad_norm": 0.5315167109100908,
"learning_rate": 3.949046761272736e-05,
"loss": 1.5997,
"step": 188
},
{
"epoch": 1.010731052984574,
"eval_loss": 2.160255193710327,
"eval_runtime": 403.1754,
"eval_samples_per_second": 6.322,
"eval_steps_per_second": 0.791,
"step": 188
},
{
"epoch": 1.0160965794768613,
"grad_norm": 1.034163349005432,
"learning_rate": 3.9369685255360175e-05,
"loss": 1.6376,
"step": 189
},
{
"epoch": 1.0214621059691482,
"grad_norm": 0.5297370695526916,
"learning_rate": 3.924839991107229e-05,
"loss": 1.5196,
"step": 190
},
{
"epoch": 1.0268276324614354,
"grad_norm": 0.49978902813452,
"learning_rate": 3.9126615825255364e-05,
"loss": 1.4628,
"step": 191
},
{
"epoch": 1.0321931589537223,
"grad_norm": 0.5207450843338,
"learning_rate": 3.900433726075865e-05,
"loss": 1.5808,
"step": 192
},
{
"epoch": 1.0375586854460095,
"grad_norm": 0.5502855697944308,
"learning_rate": 3.888156849773985e-05,
"loss": 1.5445,
"step": 193
},
{
"epoch": 1.0429242119382964,
"grad_norm": 0.5175112784168553,
"learning_rate": 3.875831383351519e-05,
"loss": 1.6456,
"step": 194
},
{
"epoch": 1.0482897384305836,
"grad_norm": 0.4649959629509902,
"learning_rate": 3.863457758240912e-05,
"loss": 1.5508,
"step": 195
},
{
"epoch": 1.0536552649228705,
"grad_norm": 0.48136239187851326,
"learning_rate": 3.851036407560319e-05,
"loss": 1.4848,
"step": 196
},
{
"epoch": 1.0590207914151577,
"grad_norm": 0.48389312973032594,
"learning_rate": 3.838567766098452e-05,
"loss": 1.5123,
"step": 197
},
{
"epoch": 1.0643863179074446,
"grad_norm": 0.45180450785183063,
"learning_rate": 3.826052270299356e-05,
"loss": 1.5807,
"step": 198
},
{
"epoch": 1.0697518443997318,
"grad_norm": 0.48079670948977926,
"learning_rate": 3.813490358247137e-05,
"loss": 1.5426,
"step": 199
},
{
"epoch": 1.0751173708920188,
"grad_norm": 0.510291013526601,
"learning_rate": 3.800882469650621e-05,
"loss": 1.5845,
"step": 200
},
{
"epoch": 1.080482897384306,
"grad_norm": 0.41649615075440893,
"learning_rate": 3.78822904582797e-05,
"loss": 1.5003,
"step": 201
},
{
"epoch": 1.0858484238765929,
"grad_norm": 0.4499681852447958,
"learning_rate": 3.7755305296912276e-05,
"loss": 1.4751,
"step": 202
},
{
"epoch": 1.09121395036888,
"grad_norm": 0.5088284534955508,
"learning_rate": 3.762787365730821e-05,
"loss": 1.5982,
"step": 203
},
{
"epoch": 1.096579476861167,
"grad_norm": 0.3929831451008196,
"learning_rate": 3.7500000000000003e-05,
"loss": 1.4991,
"step": 204
},
{
"epoch": 1.1019450033534541,
"grad_norm": 0.41107315409379536,
"learning_rate": 3.7371688800992235e-05,
"loss": 1.4837,
"step": 205
},
{
"epoch": 1.107310529845741,
"grad_norm": 0.415588978346341,
"learning_rate": 3.7242944551604914e-05,
"loss": 1.5345,
"step": 206
},
{
"epoch": 1.1126760563380282,
"grad_norm": 0.4044749735529877,
"learning_rate": 3.711377175831626e-05,
"loss": 1.4545,
"step": 207
},
{
"epoch": 1.1180415828303152,
"grad_norm": 0.44865460732176465,
"learning_rate": 3.698417494260494e-05,
"loss": 1.6391,
"step": 208
},
{
"epoch": 1.1234071093226023,
"grad_norm": 0.4479424002899852,
"learning_rate": 3.685415864079185e-05,
"loss": 1.508,
"step": 209
},
{
"epoch": 1.1287726358148893,
"grad_norm": 0.429927217929734,
"learning_rate": 3.6723727403881284e-05,
"loss": 1.5501,
"step": 210
},
{
"epoch": 1.1341381623071765,
"grad_norm": 0.5044115066105113,
"learning_rate": 3.659288579740163e-05,
"loss": 1.5615,
"step": 211
},
{
"epoch": 1.1395036887994634,
"grad_norm": 0.424575600191598,
"learning_rate": 3.646163840124561e-05,
"loss": 1.4835,
"step": 212
},
{
"epoch": 1.1448692152917506,
"grad_norm": 0.43480586168977786,
"learning_rate": 3.632998980950993e-05,
"loss": 1.5371,
"step": 213
},
{
"epoch": 1.1502347417840375,
"grad_norm": 0.40322150224409975,
"learning_rate": 3.619794463033447e-05,
"loss": 1.5376,
"step": 214
},
{
"epoch": 1.1556002682763247,
"grad_norm": 0.42637706485283583,
"learning_rate": 3.6065507485741e-05,
"loss": 1.5631,
"step": 215
},
{
"epoch": 1.1609657947686116,
"grad_norm": 0.4259143981435151,
"learning_rate": 3.593268301147139e-05,
"loss": 1.5741,
"step": 216
},
{
"epoch": 1.1663313212608988,
"grad_norm": 0.45826011738708783,
"learning_rate": 3.5799475856825326e-05,
"loss": 1.5298,
"step": 217
},
{
"epoch": 1.1716968477531857,
"grad_norm": 0.4030875431169561,
"learning_rate": 3.566589068449761e-05,
"loss": 1.5574,
"step": 218
},
{
"epoch": 1.1770623742454729,
"grad_norm": 0.409604426981432,
"learning_rate": 3.5531932170414896e-05,
"loss": 1.4859,
"step": 219
},
{
"epoch": 1.1824279007377598,
"grad_norm": 0.37439697455127857,
"learning_rate": 3.539760500357207e-05,
"loss": 1.5221,
"step": 220
},
{
"epoch": 1.187793427230047,
"grad_norm": 0.3600455315943667,
"learning_rate": 3.5262913885868066e-05,
"loss": 1.479,
"step": 221
},
{
"epoch": 1.193158953722334,
"grad_norm": 0.39515964344444054,
"learning_rate": 3.512786353194134e-05,
"loss": 1.6002,
"step": 222
},
{
"epoch": 1.198524480214621,
"grad_norm": 0.3903215407425732,
"learning_rate": 3.49924586690048e-05,
"loss": 1.5627,
"step": 223
},
{
"epoch": 1.203890006706908,
"grad_norm": 0.3863928913911809,
"learning_rate": 3.485670403668036e-05,
"loss": 1.4894,
"step": 224
},
{
"epoch": 1.2092555331991952,
"grad_norm": 0.39120322752660747,
"learning_rate": 3.472060438683302e-05,
"loss": 1.5576,
"step": 225
},
{
"epoch": 1.2146210596914822,
"grad_norm": 0.4012276022035902,
"learning_rate": 3.4584164483404544e-05,
"loss": 1.5661,
"step": 226
},
{
"epoch": 1.2199865861837693,
"grad_norm": 0.359333036872216,
"learning_rate": 3.444738910224671e-05,
"loss": 1.6043,
"step": 227
},
{
"epoch": 1.2253521126760563,
"grad_norm": 0.4169926695855886,
"learning_rate": 3.431028303095415e-05,
"loss": 1.5162,
"step": 228
},
{
"epoch": 1.2307176391683434,
"grad_norm": 0.3666243644496504,
"learning_rate": 3.417285106869673e-05,
"loss": 1.5182,
"step": 229
},
{
"epoch": 1.2360831656606304,
"grad_norm": 0.3732640706992433,
"learning_rate": 3.403509802605159e-05,
"loss": 1.6136,
"step": 230
},
{
"epoch": 1.2414486921529175,
"grad_norm": 0.38533992909636827,
"learning_rate": 3.389702872483477e-05,
"loss": 1.5763,
"step": 231
},
{
"epoch": 1.2468142186452045,
"grad_norm": 0.3621197211070038,
"learning_rate": 3.3758647997932417e-05,
"loss": 1.5977,
"step": 232
},
{
"epoch": 1.2521797451374916,
"grad_norm": 0.39624430912627595,
"learning_rate": 3.361996068913159e-05,
"loss": 1.5794,
"step": 233
},
{
"epoch": 1.2575452716297786,
"grad_norm": 0.38404789088021773,
"learning_rate": 3.348097165295076e-05,
"loss": 1.5633,
"step": 234
},
{
"epoch": 1.2629107981220657,
"grad_norm": 0.35832957307552504,
"learning_rate": 3.3341685754469856e-05,
"loss": 1.5134,
"step": 235
},
{
"epoch": 1.2629107981220657,
"eval_loss": 2.15519642829895,
"eval_runtime": 403.4218,
"eval_samples_per_second": 6.318,
"eval_steps_per_second": 0.791,
"step": 235
},
{
"epoch": 1.268276324614353,
"grad_norm": 0.34366404066845024,
"learning_rate": 3.320210786915997e-05,
"loss": 1.4825,
"step": 236
},
{
"epoch": 1.2736418511066399,
"grad_norm": 0.45636968737684863,
"learning_rate": 3.3062242882712724e-05,
"loss": 1.5146,
"step": 237
},
{
"epoch": 1.2790073775989268,
"grad_norm": 0.3804000236233779,
"learning_rate": 3.2922095690869224e-05,
"loss": 1.5501,
"step": 238
},
{
"epoch": 1.284372904091214,
"grad_norm": 0.3699794774550946,
"learning_rate": 3.278167119924872e-05,
"loss": 1.4956,
"step": 239
},
{
"epoch": 1.2897384305835011,
"grad_norm": 0.38134052511434596,
"learning_rate": 3.2640974323176846e-05,
"loss": 1.4926,
"step": 240
},
{
"epoch": 1.295103957075788,
"grad_norm": 0.4001577679575819,
"learning_rate": 3.2500009987513655e-05,
"loss": 1.5339,
"step": 241
},
{
"epoch": 1.300469483568075,
"grad_norm": 0.33599496371938614,
"learning_rate": 3.235878312648112e-05,
"loss": 1.3329,
"step": 242
},
{
"epoch": 1.3058350100603622,
"grad_norm": 0.3780832169945631,
"learning_rate": 3.2217298683490525e-05,
"loss": 1.5711,
"step": 243
},
{
"epoch": 1.3112005365526493,
"grad_norm": 0.37979912787014874,
"learning_rate": 3.207556161096935e-05,
"loss": 1.559,
"step": 244
},
{
"epoch": 1.3165660630449363,
"grad_norm": 0.34013678855872914,
"learning_rate": 3.193357687018798e-05,
"loss": 1.5112,
"step": 245
},
{
"epoch": 1.3219315895372232,
"grad_norm": 0.3761770494247479,
"learning_rate": 3.179134943108597e-05,
"loss": 1.5195,
"step": 246
},
{
"epoch": 1.3272971160295104,
"grad_norm": 0.36046502801340735,
"learning_rate": 3.164888427209818e-05,
"loss": 1.4648,
"step": 247
},
{
"epoch": 1.3326626425217976,
"grad_norm": 0.4163102856970103,
"learning_rate": 3.150618637998041e-05,
"loss": 1.6488,
"step": 248
},
{
"epoch": 1.3380281690140845,
"grad_norm": 0.37250576413738173,
"learning_rate": 3.136326074963494e-05,
"loss": 1.5966,
"step": 249
},
{
"epoch": 1.3433936955063714,
"grad_norm": 0.3717314773044194,
"learning_rate": 3.122011238393562e-05,
"loss": 1.4555,
"step": 250
},
{
"epoch": 1.3487592219986586,
"grad_norm": 0.4233051858413458,
"learning_rate": 3.1076746293552786e-05,
"loss": 1.5931,
"step": 251
},
{
"epoch": 1.3541247484909458,
"grad_norm": 0.3538900779699424,
"learning_rate": 3.093316749677788e-05,
"loss": 1.5041,
"step": 252
},
{
"epoch": 1.3594902749832327,
"grad_norm": 0.4239252371757693,
"learning_rate": 3.078938101934773e-05,
"loss": 1.5986,
"step": 253
},
{
"epoch": 1.3648558014755197,
"grad_norm": 0.40455165331708126,
"learning_rate": 3.064539189426874e-05,
"loss": 1.5956,
"step": 254
},
{
"epoch": 1.3702213279678068,
"grad_norm": 0.35682161127333073,
"learning_rate": 3.050120516164062e-05,
"loss": 1.4983,
"step": 255
},
{
"epoch": 1.375586854460094,
"grad_norm": 0.366685178903664,
"learning_rate": 3.0356825868480017e-05,
"loss": 1.6384,
"step": 256
},
{
"epoch": 1.380952380952381,
"grad_norm": 0.3615423341669086,
"learning_rate": 3.0212259068543837e-05,
"loss": 1.4734,
"step": 257
},
{
"epoch": 1.3863179074446679,
"grad_norm": 0.3469703962175405,
"learning_rate": 3.006750982215234e-05,
"loss": 1.4386,
"step": 258
},
{
"epoch": 1.391683433936955,
"grad_norm": 0.3757992095280783,
"learning_rate": 2.9922583196012037e-05,
"loss": 1.5479,
"step": 259
},
{
"epoch": 1.3970489604292422,
"grad_norm": 0.3696529318558144,
"learning_rate": 2.9777484263038306e-05,
"loss": 1.4613,
"step": 260
},
{
"epoch": 1.4024144869215291,
"grad_norm": 0.36763443681114544,
"learning_rate": 2.9632218102177862e-05,
"loss": 1.4707,
"step": 261
},
{
"epoch": 1.4077800134138163,
"grad_norm": 0.3330227260573098,
"learning_rate": 2.9486789798230917e-05,
"loss": 1.5196,
"step": 262
},
{
"epoch": 1.4131455399061033,
"grad_norm": 0.3401075705229897,
"learning_rate": 2.9341204441673266e-05,
"loss": 1.5713,
"step": 263
},
{
"epoch": 1.4185110663983904,
"grad_norm": 0.38589937071982083,
"learning_rate": 2.9195467128478044e-05,
"loss": 1.5658,
"step": 264
},
{
"epoch": 1.4238765928906774,
"grad_norm": 0.36187499541792045,
"learning_rate": 2.9049582959937392e-05,
"loss": 1.5645,
"step": 265
},
{
"epoch": 1.4292421193829645,
"grad_norm": 0.3727884692802974,
"learning_rate": 2.8903557042483887e-05,
"loss": 1.5195,
"step": 266
},
{
"epoch": 1.4346076458752515,
"grad_norm": 0.3423144867083204,
"learning_rate": 2.875739448751176e-05,
"loss": 1.4897,
"step": 267
},
{
"epoch": 1.4399731723675386,
"grad_norm": 0.3809610047831763,
"learning_rate": 2.8611100411198037e-05,
"loss": 1.6107,
"step": 268
},
{
"epoch": 1.4453386988598256,
"grad_norm": 0.3466660038008262,
"learning_rate": 2.8464679934323424e-05,
"loss": 1.5203,
"step": 269
},
{
"epoch": 1.4507042253521127,
"grad_norm": 0.3858219936970091,
"learning_rate": 2.8318138182093052e-05,
"loss": 1.5211,
"step": 270
},
{
"epoch": 1.4560697518443997,
"grad_norm": 0.3457606288538911,
"learning_rate": 2.8171480283957118e-05,
"loss": 1.4825,
"step": 271
},
{
"epoch": 1.4614352783366868,
"grad_norm": 0.4097169980235964,
"learning_rate": 2.80247113734313e-05,
"loss": 1.5276,
"step": 272
},
{
"epoch": 1.4668008048289738,
"grad_norm": 0.39937404860513825,
"learning_rate": 2.7877836587917072e-05,
"loss": 1.5022,
"step": 273
},
{
"epoch": 1.472166331321261,
"grad_norm": 0.36070057912441533,
"learning_rate": 2.773086106852192e-05,
"loss": 1.4587,
"step": 274
},
{
"epoch": 1.477531857813548,
"grad_norm": 0.4013616820391148,
"learning_rate": 2.7583789959879303e-05,
"loss": 1.5908,
"step": 275
},
{
"epoch": 1.482897384305835,
"grad_norm": 0.4150531718746046,
"learning_rate": 2.7436628409968664e-05,
"loss": 1.4511,
"step": 276
},
{
"epoch": 1.488262910798122,
"grad_norm": 0.38578622280087393,
"learning_rate": 2.728938156993517e-05,
"loss": 1.5407,
"step": 277
},
{
"epoch": 1.4936284372904092,
"grad_norm": 0.3844783258031986,
"learning_rate": 2.7142054593909422e-05,
"loss": 1.5349,
"step": 278
},
{
"epoch": 1.4989939637826961,
"grad_norm": 0.4144431176495637,
"learning_rate": 2.6994652638827078e-05,
"loss": 1.602,
"step": 279
},
{
"epoch": 1.5043594902749833,
"grad_norm": 0.402711914294424,
"learning_rate": 2.6847180864248283e-05,
"loss": 1.5902,
"step": 280
},
{
"epoch": 1.5097250167672702,
"grad_norm": 0.3378799881196754,
"learning_rate": 2.6699644432177112e-05,
"loss": 1.5514,
"step": 281
},
{
"epoch": 1.5150905432595574,
"grad_norm": 0.4113537256498264,
"learning_rate": 2.655204850688085e-05,
"loss": 1.4614,
"step": 282
},
{
"epoch": 1.5150905432595574,
"eval_loss": 2.1373159885406494,
"eval_runtime": 403.7482,
"eval_samples_per_second": 6.313,
"eval_steps_per_second": 0.79,
"step": 282
},
{
"epoch": 1.5204560697518446,
"grad_norm": 0.3645339916373801,
"learning_rate": 2.6404398254709284e-05,
"loss": 1.4716,
"step": 283
},
{
"epoch": 1.5258215962441315,
"grad_norm": 0.3662916121671904,
"learning_rate": 2.625669884391377e-05,
"loss": 1.5145,
"step": 284
},
{
"epoch": 1.5311871227364184,
"grad_norm": 0.4062086365442157,
"learning_rate": 2.610895544446641e-05,
"loss": 1.5513,
"step": 285
},
{
"epoch": 1.5365526492287056,
"grad_norm": 0.3693755738805308,
"learning_rate": 2.596117322787907e-05,
"loss": 1.5498,
"step": 286
},
{
"epoch": 1.5419181757209928,
"grad_norm": 0.3589026792727439,
"learning_rate": 2.5813357367022305e-05,
"loss": 1.5211,
"step": 287
},
{
"epoch": 1.5472837022132797,
"grad_norm": 0.38017444587432575,
"learning_rate": 2.566551303594437e-05,
"loss": 1.5342,
"step": 288
},
{
"epoch": 1.5526492287055667,
"grad_norm": 0.3537300073106448,
"learning_rate": 2.551764540969005e-05,
"loss": 1.5109,
"step": 289
},
{
"epoch": 1.5580147551978538,
"grad_norm": 0.3752836981822812,
"learning_rate": 2.5369759664119537e-05,
"loss": 1.5015,
"step": 290
},
{
"epoch": 1.563380281690141,
"grad_norm": 0.3810876452684417,
"learning_rate": 2.5221860975727275e-05,
"loss": 1.5686,
"step": 291
},
{
"epoch": 1.568745808182428,
"grad_norm": 0.3688966465073692,
"learning_rate": 2.5073954521460745e-05,
"loss": 1.5666,
"step": 292
},
{
"epoch": 1.5741113346747149,
"grad_norm": 0.3740373962224673,
"learning_rate": 2.4926045478539257e-05,
"loss": 1.5541,
"step": 293
},
{
"epoch": 1.579476861167002,
"grad_norm": 0.4415523054571514,
"learning_rate": 2.4778139024272724e-05,
"loss": 1.62,
"step": 294
},
{
"epoch": 1.5848423876592892,
"grad_norm": 0.3291219023047839,
"learning_rate": 2.4630240335880462e-05,
"loss": 1.5,
"step": 295
},
{
"epoch": 1.5902079141515761,
"grad_norm": 0.39622699819562734,
"learning_rate": 2.4482354590309962e-05,
"loss": 1.5358,
"step": 296
},
{
"epoch": 1.595573440643863,
"grad_norm": 0.35147156988608064,
"learning_rate": 2.433448696405563e-05,
"loss": 1.469,
"step": 297
},
{
"epoch": 1.6009389671361502,
"grad_norm": 0.3458967842322882,
"learning_rate": 2.4186642632977697e-05,
"loss": 1.5368,
"step": 298
},
{
"epoch": 1.6063044936284374,
"grad_norm": 0.45919555248078225,
"learning_rate": 2.4038826772120932e-05,
"loss": 1.6126,
"step": 299
},
{
"epoch": 1.6116700201207244,
"grad_norm": 0.397298560652149,
"learning_rate": 2.3891044555533588e-05,
"loss": 1.5273,
"step": 300
},
{
"epoch": 1.6170355466130113,
"grad_norm": 0.33950881912700964,
"learning_rate": 2.3743301156086244e-05,
"loss": 1.5844,
"step": 301
},
{
"epoch": 1.6224010731052985,
"grad_norm": 0.3685147478174407,
"learning_rate": 2.359560174529073e-05,
"loss": 1.5306,
"step": 302
},
{
"epoch": 1.6277665995975856,
"grad_norm": 0.3668847123777601,
"learning_rate": 2.3447951493119152e-05,
"loss": 1.5114,
"step": 303
},
{
"epoch": 1.6331321260898726,
"grad_norm": 0.3532238202377609,
"learning_rate": 2.3300355567822897e-05,
"loss": 1.5568,
"step": 304
},
{
"epoch": 1.6384976525821595,
"grad_norm": 0.3810645993344603,
"learning_rate": 2.3152819135751722e-05,
"loss": 1.5121,
"step": 305
},
{
"epoch": 1.6438631790744467,
"grad_norm": 0.37909917013988087,
"learning_rate": 2.300534736117292e-05,
"loss": 1.4904,
"step": 306
},
{
"epoch": 1.6492287055667338,
"grad_norm": 0.38885366059907706,
"learning_rate": 2.285794540609058e-05,
"loss": 1.5852,
"step": 307
},
{
"epoch": 1.6545942320590208,
"grad_norm": 0.3657756765079751,
"learning_rate": 2.2710618430064843e-05,
"loss": 1.5529,
"step": 308
},
{
"epoch": 1.6599597585513077,
"grad_norm": 0.3397163303578702,
"learning_rate": 2.256337159003134e-05,
"loss": 1.4903,
"step": 309
},
{
"epoch": 1.665325285043595,
"grad_norm": 0.39777082501030225,
"learning_rate": 2.2416210040120703e-05,
"loss": 1.5159,
"step": 310
},
{
"epoch": 1.670690811535882,
"grad_norm": 0.3658755281773237,
"learning_rate": 2.2269138931478084e-05,
"loss": 1.5848,
"step": 311
},
{
"epoch": 1.676056338028169,
"grad_norm": 0.42665798217406753,
"learning_rate": 2.2122163412082927e-05,
"loss": 1.6133,
"step": 312
},
{
"epoch": 1.681421864520456,
"grad_norm": 0.40969972656814363,
"learning_rate": 2.1975288626568713e-05,
"loss": 1.6264,
"step": 313
},
{
"epoch": 1.686787391012743,
"grad_norm": 0.3283642172163527,
"learning_rate": 2.1828519716042888e-05,
"loss": 1.4812,
"step": 314
},
{
"epoch": 1.6921529175050303,
"grad_norm": 0.4058354316614303,
"learning_rate": 2.1681861817906954e-05,
"loss": 1.4827,
"step": 315
},
{
"epoch": 1.6975184439973172,
"grad_norm": 0.3466585039953347,
"learning_rate": 2.153532006567658e-05,
"loss": 1.5768,
"step": 316
},
{
"epoch": 1.7028839704896042,
"grad_norm": 0.33329283442783797,
"learning_rate": 2.1388899588801965e-05,
"loss": 1.577,
"step": 317
},
{
"epoch": 1.7082494969818913,
"grad_norm": 0.3513005805125947,
"learning_rate": 2.1242605512488248e-05,
"loss": 1.5835,
"step": 318
},
{
"epoch": 1.7136150234741785,
"grad_norm": 0.3294934417641562,
"learning_rate": 2.109644295751612e-05,
"loss": 1.4758,
"step": 319
},
{
"epoch": 1.7189805499664654,
"grad_norm": 0.33415233723527016,
"learning_rate": 2.095041704006261e-05,
"loss": 1.4853,
"step": 320
},
{
"epoch": 1.7243460764587524,
"grad_norm": 0.3283653112019817,
"learning_rate": 2.080453287152196e-05,
"loss": 1.4554,
"step": 321
},
{
"epoch": 1.7297116029510395,
"grad_norm": 0.32083254638693504,
"learning_rate": 2.0658795558326743e-05,
"loss": 1.5284,
"step": 322
},
{
"epoch": 1.7350771294433267,
"grad_norm": 0.37794249417672565,
"learning_rate": 2.0513210201769085e-05,
"loss": 1.5521,
"step": 323
},
{
"epoch": 1.7404426559356136,
"grad_norm": 0.3431657733658665,
"learning_rate": 2.0367781897822147e-05,
"loss": 1.5067,
"step": 324
},
{
"epoch": 1.7458081824279006,
"grad_norm": 0.33934139214309345,
"learning_rate": 2.0222515736961696e-05,
"loss": 1.4801,
"step": 325
},
{
"epoch": 1.7511737089201878,
"grad_norm": 0.3795356143714155,
"learning_rate": 2.0077416803987965e-05,
"loss": 1.6264,
"step": 326
},
{
"epoch": 1.756539235412475,
"grad_norm": 0.3507795265759812,
"learning_rate": 1.993249017784766e-05,
"loss": 1.4967,
"step": 327
},
{
"epoch": 1.7619047619047619,
"grad_norm": 0.3648489232026123,
"learning_rate": 1.9787740931456165e-05,
"loss": 1.5561,
"step": 328
},
{
"epoch": 1.7672702883970488,
"grad_norm": 0.39626816117060937,
"learning_rate": 1.9643174131519986e-05,
"loss": 1.6362,
"step": 329
},
{
"epoch": 1.7672702883970488,
"eval_loss": 2.1242423057556152,
"eval_runtime": 403.1029,
"eval_samples_per_second": 6.323,
"eval_steps_per_second": 0.791,
"step": 329
},
{
"epoch": 1.772635814889336,
"grad_norm": 0.35313890178416246,
"learning_rate": 1.949879483835939e-05,
"loss": 1.5147,
"step": 330
},
{
"epoch": 1.7780013413816231,
"grad_norm": 0.3269413977690858,
"learning_rate": 1.935460810573127e-05,
"loss": 1.5032,
"step": 331
},
{
"epoch": 1.78336686787391,
"grad_norm": 0.3521215800086969,
"learning_rate": 1.9210618980652277e-05,
"loss": 1.5734,
"step": 332
},
{
"epoch": 1.788732394366197,
"grad_norm": 0.3813859592452455,
"learning_rate": 1.9066832503222128e-05,
"loss": 1.4488,
"step": 333
},
{
"epoch": 1.7940979208584842,
"grad_norm": 0.31690523060028536,
"learning_rate": 1.892325370644721e-05,
"loss": 1.432,
"step": 334
},
{
"epoch": 1.7994634473507714,
"grad_norm": 0.3364023400710833,
"learning_rate": 1.8779887616064383e-05,
"loss": 1.4871,
"step": 335
},
{
"epoch": 1.8048289738430583,
"grad_norm": 0.3350534986332929,
"learning_rate": 1.863673925036506e-05,
"loss": 1.5055,
"step": 336
},
{
"epoch": 1.8101945003353455,
"grad_norm": 0.33398900975140067,
"learning_rate": 1.8493813620019594e-05,
"loss": 1.5005,
"step": 337
},
{
"epoch": 1.8155600268276326,
"grad_norm": 0.31906353499540524,
"learning_rate": 1.835111572790183e-05,
"loss": 1.4626,
"step": 338
},
{
"epoch": 1.8209255533199196,
"grad_norm": 0.34716461667960885,
"learning_rate": 1.8208650568914033e-05,
"loss": 1.509,
"step": 339
},
{
"epoch": 1.8262910798122065,
"grad_norm": 0.34582309441917697,
"learning_rate": 1.8066423129812027e-05,
"loss": 1.4824,
"step": 340
},
{
"epoch": 1.8316566063044937,
"grad_norm": 0.32325626999744855,
"learning_rate": 1.792443838903065e-05,
"loss": 1.4303,
"step": 341
},
{
"epoch": 1.8370221327967808,
"grad_norm": 0.32008576746121775,
"learning_rate": 1.778270131650948e-05,
"loss": 1.6245,
"step": 342
},
{
"epoch": 1.8423876592890678,
"grad_norm": 0.37457992451894306,
"learning_rate": 1.7641216873518878e-05,
"loss": 1.4618,
"step": 343
},
{
"epoch": 1.8477531857813547,
"grad_norm": 0.35750145316045384,
"learning_rate": 1.749999001248635e-05,
"loss": 1.5959,
"step": 344
},
{
"epoch": 1.8531187122736419,
"grad_norm": 0.3306845841796733,
"learning_rate": 1.735902567682315e-05,
"loss": 1.5816,
"step": 345
},
{
"epoch": 1.858484238765929,
"grad_norm": 0.38970827119176255,
"learning_rate": 1.7218328800751288e-05,
"loss": 1.5529,
"step": 346
},
{
"epoch": 1.863849765258216,
"grad_norm": 0.3217477415509899,
"learning_rate": 1.7077904309130782e-05,
"loss": 1.5559,
"step": 347
},
{
"epoch": 1.869215291750503,
"grad_norm": 0.36288548537121584,
"learning_rate": 1.6937757117287278e-05,
"loss": 1.4984,
"step": 348
},
{
"epoch": 1.87458081824279,
"grad_norm": 0.3460179189739247,
"learning_rate": 1.6797892130840036e-05,
"loss": 1.5385,
"step": 349
},
{
"epoch": 1.8799463447350773,
"grad_norm": 0.31554481787674976,
"learning_rate": 1.665831424553015e-05,
"loss": 1.504,
"step": 350
},
{
"epoch": 1.8853118712273642,
"grad_norm": 0.30549498763968225,
"learning_rate": 1.651902834704924e-05,
"loss": 1.5228,
"step": 351
},
{
"epoch": 1.8906773977196512,
"grad_norm": 0.3277064150411407,
"learning_rate": 1.6380039310868416e-05,
"loss": 1.4852,
"step": 352
},
{
"epoch": 1.8960429242119383,
"grad_norm": 0.31968263622304366,
"learning_rate": 1.624135200206759e-05,
"loss": 1.4611,
"step": 353
},
{
"epoch": 1.9014084507042255,
"grad_norm": 0.32902733969646153,
"learning_rate": 1.6102971275165228e-05,
"loss": 1.4833,
"step": 354
},
{
"epoch": 1.9067739771965124,
"grad_norm": 0.3299428742675132,
"learning_rate": 1.596490197394841e-05,
"loss": 1.4439,
"step": 355
},
{
"epoch": 1.9121395036887994,
"grad_norm": 0.34500427705894593,
"learning_rate": 1.5827148931303277e-05,
"loss": 1.5616,
"step": 356
},
{
"epoch": 1.9175050301810865,
"grad_norm": 0.32458276939765035,
"learning_rate": 1.5689716969045848e-05,
"loss": 1.4334,
"step": 357
},
{
"epoch": 1.9228705566733737,
"grad_norm": 0.3342429982300284,
"learning_rate": 1.5552610897753292e-05,
"loss": 1.5311,
"step": 358
},
{
"epoch": 1.9282360831656606,
"grad_norm": 0.3483438918440925,
"learning_rate": 1.5415835516595465e-05,
"loss": 1.4642,
"step": 359
},
{
"epoch": 1.9336016096579476,
"grad_norm": 0.324692472057597,
"learning_rate": 1.5279395613166986e-05,
"loss": 1.5336,
"step": 360
},
{
"epoch": 1.9389671361502347,
"grad_norm": 0.4011901483013197,
"learning_rate": 1.5143295963319643e-05,
"loss": 1.5634,
"step": 361
},
{
"epoch": 1.944332662642522,
"grad_norm": 0.3791986336861211,
"learning_rate": 1.5007541330995197e-05,
"loss": 1.5722,
"step": 362
},
{
"epoch": 1.9496981891348089,
"grad_norm": 0.3318324328550346,
"learning_rate": 1.4872136468058661e-05,
"loss": 1.6214,
"step": 363
},
{
"epoch": 1.9550637156270958,
"grad_norm": 0.3865281831290857,
"learning_rate": 1.4737086114131943e-05,
"loss": 1.548,
"step": 364
},
{
"epoch": 1.960429242119383,
"grad_norm": 0.3317269613634914,
"learning_rate": 1.4602394996427942e-05,
"loss": 1.5024,
"step": 365
},
{
"epoch": 1.9657947686116701,
"grad_norm": 0.36387655313468187,
"learning_rate": 1.4468067829585108e-05,
"loss": 1.5256,
"step": 366
},
{
"epoch": 1.971160295103957,
"grad_norm": 0.3494371010379711,
"learning_rate": 1.4334109315502395e-05,
"loss": 1.5559,
"step": 367
},
{
"epoch": 1.976525821596244,
"grad_norm": 0.36135083308665883,
"learning_rate": 1.4200524143174677e-05,
"loss": 1.5669,
"step": 368
},
{
"epoch": 1.9818913480885312,
"grad_norm": 0.34824673781683024,
"learning_rate": 1.4067316988528617e-05,
"loss": 1.6642,
"step": 369
},
{
"epoch": 1.9872568745808183,
"grad_norm": 0.36134664903507996,
"learning_rate": 1.3934492514259003e-05,
"loss": 1.5003,
"step": 370
},
{
"epoch": 1.9926224010731053,
"grad_norm": 0.3315306163140975,
"learning_rate": 1.3802055369665534e-05,
"loss": 1.3959,
"step": 371
},
{
"epoch": 1.9979879275653922,
"grad_norm": 0.3179771784326003,
"learning_rate": 1.3670010190490073e-05,
"loss": 1.5285,
"step": 372
},
{
"epoch": 2.005365526492287,
"grad_norm": 0.9915595865019317,
"learning_rate": 1.3538361598754384e-05,
"loss": 2.456,
"step": 373
},
{
"epoch": 2.0107310529845743,
"grad_norm": 0.6291566214590382,
"learning_rate": 1.3407114202598369e-05,
"loss": 1.1708,
"step": 374
},
{
"epoch": 2.0160965794768613,
"grad_norm": 0.5593895609190137,
"learning_rate": 1.327627259611873e-05,
"loss": 1.1864,
"step": 375
},
{
"epoch": 2.021462105969148,
"grad_norm": 0.4205563210949494,
"learning_rate": 1.314584135920815e-05,
"loss": 1.1374,
"step": 376
},
{
"epoch": 2.021462105969148,
"eval_loss": 2.232717752456665,
"eval_runtime": 403.1787,
"eval_samples_per_second": 6.322,
"eval_steps_per_second": 0.791,
"step": 376
},
{
"epoch": 2.026827632461435,
"grad_norm": 0.7137961053387198,
"learning_rate": 1.301582505739506e-05,
"loss": 1.2243,
"step": 377
},
{
"epoch": 2.0321931589537225,
"grad_norm": 0.7110032999802158,
"learning_rate": 1.2886228241683749e-05,
"loss": 1.1914,
"step": 378
},
{
"epoch": 2.0375586854460095,
"grad_norm": 0.6070986317073753,
"learning_rate": 1.2757055448395092e-05,
"loss": 1.134,
"step": 379
},
{
"epoch": 2.0429242119382964,
"grad_norm": 0.4599660060017544,
"learning_rate": 1.2628311199007764e-05,
"loss": 1.1466,
"step": 380
},
{
"epoch": 2.0482897384305834,
"grad_norm": 0.48013581218581647,
"learning_rate": 1.2500000000000006e-05,
"loss": 1.2032,
"step": 381
},
{
"epoch": 2.0536552649228708,
"grad_norm": 0.47190986939626134,
"learning_rate": 1.2372126342691798e-05,
"loss": 1.1103,
"step": 382
},
{
"epoch": 2.0590207914151577,
"grad_norm": 0.4474749587522701,
"learning_rate": 1.2244694703087728e-05,
"loss": 1.1502,
"step": 383
},
{
"epoch": 2.0643863179074446,
"grad_norm": 0.44999362516877806,
"learning_rate": 1.2117709541720306e-05,
"loss": 1.1528,
"step": 384
},
{
"epoch": 2.0697518443997316,
"grad_norm": 0.45904376696237503,
"learning_rate": 1.1991175303493793e-05,
"loss": 1.2056,
"step": 385
},
{
"epoch": 2.075117370892019,
"grad_norm": 0.4344382184186286,
"learning_rate": 1.1865096417528635e-05,
"loss": 1.1107,
"step": 386
},
{
"epoch": 2.080482897384306,
"grad_norm": 0.3972755677260206,
"learning_rate": 1.173947729700644e-05,
"loss": 1.1499,
"step": 387
},
{
"epoch": 2.085848423876593,
"grad_norm": 0.3950446642940489,
"learning_rate": 1.1614322339015484e-05,
"loss": 1.1557,
"step": 388
},
{
"epoch": 2.09121395036888,
"grad_norm": 0.4447326020481244,
"learning_rate": 1.1489635924396817e-05,
"loss": 1.1555,
"step": 389
},
{
"epoch": 2.096579476861167,
"grad_norm": 0.419897228759811,
"learning_rate": 1.1365422417590878e-05,
"loss": 1.1756,
"step": 390
},
{
"epoch": 2.101945003353454,
"grad_norm": 0.41249904086804146,
"learning_rate": 1.1241686166484805e-05,
"loss": 1.1706,
"step": 391
},
{
"epoch": 2.107310529845741,
"grad_norm": 0.40775526734089124,
"learning_rate": 1.1118431502260163e-05,
"loss": 1.2389,
"step": 392
},
{
"epoch": 2.112676056338028,
"grad_norm": 0.41120972999065375,
"learning_rate": 1.0995662739241347e-05,
"loss": 1.2027,
"step": 393
},
{
"epoch": 2.1180415828303154,
"grad_norm": 0.36915354132829215,
"learning_rate": 1.0873384174744641e-05,
"loss": 1.1211,
"step": 394
},
{
"epoch": 2.1234071093226023,
"grad_norm": 0.3734288072829599,
"learning_rate": 1.0751600088927713e-05,
"loss": 1.2408,
"step": 395
},
{
"epoch": 2.1287726358148893,
"grad_norm": 0.40175568463326916,
"learning_rate": 1.063031474463983e-05,
"loss": 1.1744,
"step": 396
},
{
"epoch": 2.1341381623071762,
"grad_norm": 0.3674474877057401,
"learning_rate": 1.0509532387272641e-05,
"loss": 1.1628,
"step": 397
},
{
"epoch": 2.1395036887994636,
"grad_norm": 0.36762315854199107,
"learning_rate": 1.0389257244611602e-05,
"loss": 1.162,
"step": 398
},
{
"epoch": 2.1448692152917506,
"grad_norm": 0.3414954721940739,
"learning_rate": 1.0269493526687915e-05,
"loss": 1.182,
"step": 399
},
{
"epoch": 2.1502347417840375,
"grad_norm": 0.38061259662736135,
"learning_rate": 1.0150245425631235e-05,
"loss": 1.1561,
"step": 400
},
{
"epoch": 2.1556002682763244,
"grad_norm": 0.36884370471677075,
"learning_rate": 1.0031517115522926e-05,
"loss": 1.2107,
"step": 401
},
{
"epoch": 2.160965794768612,
"grad_norm": 0.3461756268853831,
"learning_rate": 9.913312752249903e-06,
"loss": 1.1919,
"step": 402
},
{
"epoch": 2.166331321260899,
"grad_norm": 0.3453066453102883,
"learning_rate": 9.795636473359208e-06,
"loss": 1.1735,
"step": 403
},
{
"epoch": 2.1716968477531857,
"grad_norm": 0.3585272022980633,
"learning_rate": 9.678492397913167e-06,
"loss": 1.2119,
"step": 404
},
{
"epoch": 2.1770623742454727,
"grad_norm": 0.34484629378614046,
"learning_rate": 9.561884626345205e-06,
"loss": 1.1628,
"step": 405
},
{
"epoch": 2.18242790073776,
"grad_norm": 0.340448631343189,
"learning_rate": 9.445817240316332e-06,
"loss": 1.1534,
"step": 406
},
{
"epoch": 2.187793427230047,
"grad_norm": 0.3491564049843763,
"learning_rate": 9.330294302572242e-06,
"loss": 1.1707,
"step": 407
},
{
"epoch": 2.193158953722334,
"grad_norm": 0.3374441017987585,
"learning_rate": 9.215319856801158e-06,
"loss": 1.124,
"step": 408
},
{
"epoch": 2.198524480214621,
"grad_norm": 0.33387420272144586,
"learning_rate": 9.10089792749223e-06,
"loss": 1.1547,
"step": 409
},
{
"epoch": 2.2038900067069083,
"grad_norm": 0.34377542398740535,
"learning_rate": 8.987032519794666e-06,
"loss": 1.1865,
"step": 410
},
{
"epoch": 2.209255533199195,
"grad_norm": 0.34410948368421596,
"learning_rate": 8.873727619377611e-06,
"loss": 1.1807,
"step": 411
},
{
"epoch": 2.214621059691482,
"grad_norm": 0.3454829815928742,
"learning_rate": 8.760987192290557e-06,
"loss": 1.2488,
"step": 412
},
{
"epoch": 2.219986586183769,
"grad_norm": 0.34035820590189597,
"learning_rate": 8.648815184824544e-06,
"loss": 1.1629,
"step": 413
},
{
"epoch": 2.2253521126760565,
"grad_norm": 0.35672991922416875,
"learning_rate": 8.537215523374038e-06,
"loss": 1.2238,
"step": 414
},
{
"epoch": 2.2307176391683434,
"grad_norm": 0.351148929709105,
"learning_rate": 8.426192114299484e-06,
"loss": 1.2074,
"step": 415
},
{
"epoch": 2.2360831656606304,
"grad_norm": 0.3449030405647957,
"learning_rate": 8.315748843790563e-06,
"loss": 1.188,
"step": 416
},
{
"epoch": 2.2414486921529173,
"grad_norm": 0.3432687950624104,
"learning_rate": 8.20588957773018e-06,
"loss": 1.1524,
"step": 417
},
{
"epoch": 2.2468142186452047,
"grad_norm": 0.34423459328901956,
"learning_rate": 8.096618161559142e-06,
"loss": 1.143,
"step": 418
},
{
"epoch": 2.2521797451374916,
"grad_norm": 0.32619680555917097,
"learning_rate": 7.987938420141536e-06,
"loss": 1.1641,
"step": 419
},
{
"epoch": 2.2575452716297786,
"grad_norm": 0.33552529458130786,
"learning_rate": 7.879854157630861e-06,
"loss": 1.1301,
"step": 420
},
{
"epoch": 2.262910798122066,
"grad_norm": 0.3336572936896806,
"learning_rate": 7.772369157336874e-06,
"loss": 1.119,
"step": 421
},
{
"epoch": 2.268276324614353,
"grad_norm": 0.33482717731077677,
"learning_rate": 7.665487181593145e-06,
"loss": 1.1412,
"step": 422
},
{
"epoch": 2.27364185110664,
"grad_norm": 0.31988515510077037,
"learning_rate": 7.5592119716253855e-06,
"loss": 1.2301,
"step": 423
},
{
"epoch": 2.27364185110664,
"eval_loss": 2.2243738174438477,
"eval_runtime": 403.6162,
"eval_samples_per_second": 6.315,
"eval_steps_per_second": 0.79,
"step": 423
},
{
"epoch": 2.279007377598927,
"grad_norm": 0.32212835750220664,
"learning_rate": 7.4535472474204645e-06,
"loss": 1.1127,
"step": 424
},
{
"epoch": 2.2843729040912137,
"grad_norm": 0.3541133973544153,
"learning_rate": 7.348496707596242e-06,
"loss": 1.0762,
"step": 425
},
{
"epoch": 2.289738430583501,
"grad_norm": 0.3272644781347247,
"learning_rate": 7.244064029272049e-06,
"loss": 1.1938,
"step": 426
},
{
"epoch": 2.295103957075788,
"grad_norm": 0.3248701960652722,
"learning_rate": 7.140252867939995e-06,
"loss": 1.1952,
"step": 427
},
{
"epoch": 2.300469483568075,
"grad_norm": 0.3278958191476634,
"learning_rate": 7.037066857337058e-06,
"loss": 1.2259,
"step": 428
},
{
"epoch": 2.3058350100603624,
"grad_norm": 0.33588934698731865,
"learning_rate": 6.934509609317821e-06,
"loss": 1.0948,
"step": 429
},
{
"epoch": 2.3112005365526493,
"grad_norm": 0.33096764574513354,
"learning_rate": 6.832584713728101e-06,
"loss": 1.1655,
"step": 430
},
{
"epoch": 2.3165660630449363,
"grad_norm": 0.3363375955193017,
"learning_rate": 6.7312957382792556e-06,
"loss": 1.1786,
"step": 431
},
{
"epoch": 2.3219315895372232,
"grad_norm": 0.3384101006551099,
"learning_rate": 6.6306462284233234e-06,
"loss": 1.1761,
"step": 432
},
{
"epoch": 2.32729711602951,
"grad_norm": 0.6194898145769814,
"learning_rate": 6.5306397072289114e-06,
"loss": 1.1916,
"step": 433
},
{
"epoch": 2.3326626425217976,
"grad_norm": 0.33449734643167867,
"learning_rate": 6.431279675257873e-06,
"loss": 1.0894,
"step": 434
},
{
"epoch": 2.3380281690140845,
"grad_norm": 0.3449479370773172,
"learning_rate": 6.332569610442807e-06,
"loss": 1.2021,
"step": 435
},
{
"epoch": 2.3433936955063714,
"grad_norm": 0.3324865938198681,
"learning_rate": 6.234512967965261e-06,
"loss": 1.1583,
"step": 436
},
{
"epoch": 2.348759221998659,
"grad_norm": 0.3233624144923054,
"learning_rate": 6.1371131801348434e-06,
"loss": 1.1243,
"step": 437
},
{
"epoch": 2.3541247484909458,
"grad_norm": 0.34771192845713556,
"learning_rate": 6.040373656269041e-06,
"loss": 1.1667,
"step": 438
},
{
"epoch": 2.3594902749832327,
"grad_norm": 0.3476437588037759,
"learning_rate": 5.944297782573918e-06,
"loss": 1.1192,
"step": 439
},
{
"epoch": 2.3648558014755197,
"grad_norm": 0.36436230780329204,
"learning_rate": 5.848888922025553e-06,
"loss": 1.2521,
"step": 440
},
{
"epoch": 2.3702213279678066,
"grad_norm": 0.35739290718549466,
"learning_rate": 5.75415041425234e-06,
"loss": 1.1772,
"step": 441
},
{
"epoch": 2.375586854460094,
"grad_norm": 0.3187619530515648,
"learning_rate": 5.660085575418114e-06,
"loss": 1.0855,
"step": 442
},
{
"epoch": 2.380952380952381,
"grad_norm": 0.3402178296399225,
"learning_rate": 5.566697698106024e-06,
"loss": 1.1531,
"step": 443
},
{
"epoch": 2.386317907444668,
"grad_norm": 0.3428456906688321,
"learning_rate": 5.473990051203298e-06,
"loss": 1.1289,
"step": 444
},
{
"epoch": 2.3916834339369553,
"grad_norm": 0.3251361090262083,
"learning_rate": 5.381965879786868e-06,
"loss": 1.1864,
"step": 445
},
{
"epoch": 2.397048960429242,
"grad_norm": 0.3436986211314983,
"learning_rate": 5.290628405009717e-06,
"loss": 1.0949,
"step": 446
},
{
"epoch": 2.402414486921529,
"grad_norm": 0.3350393677646101,
"learning_rate": 5.199980823988157e-06,
"loss": 1.0944,
"step": 447
},
{
"epoch": 2.407780013413816,
"grad_norm": 0.32046838027326285,
"learning_rate": 5.110026309689922e-06,
"loss": 1.1519,
"step": 448
},
{
"epoch": 2.4131455399061035,
"grad_norm": 0.33210748691525244,
"learning_rate": 5.020768010823102e-06,
"loss": 1.216,
"step": 449
},
{
"epoch": 2.4185110663983904,
"grad_norm": 0.31198379909853313,
"learning_rate": 4.932209051725914e-06,
"loss": 1.0948,
"step": 450
},
{
"epoch": 2.4238765928906774,
"grad_norm": 0.3143147465067978,
"learning_rate": 4.844352532257351e-06,
"loss": 1.0641,
"step": 451
},
{
"epoch": 2.4292421193829643,
"grad_norm": 0.3299169974610766,
"learning_rate": 4.757201527688693e-06,
"loss": 1.1331,
"step": 452
},
{
"epoch": 2.4346076458752517,
"grad_norm": 0.3240731706047221,
"learning_rate": 4.67075908859583e-06,
"loss": 1.1424,
"step": 453
},
{
"epoch": 2.4399731723675386,
"grad_norm": 0.31726341227978605,
"learning_rate": 4.585028240752498e-06,
"loss": 1.0585,
"step": 454
},
{
"epoch": 2.4453386988598256,
"grad_norm": 0.3133109530668219,
"learning_rate": 4.500011985024363e-06,
"loss": 1.1124,
"step": 455
},
{
"epoch": 2.4507042253521125,
"grad_norm": 0.32063200368226685,
"learning_rate": 4.415713297263987e-06,
"loss": 1.127,
"step": 456
},
{
"epoch": 2.4560697518444,
"grad_norm": 0.3230470474940416,
"learning_rate": 4.332135128206666e-06,
"loss": 1.164,
"step": 457
},
{
"epoch": 2.461435278336687,
"grad_norm": 0.32502523603989114,
"learning_rate": 4.2492804033671145e-06,
"loss": 1.1437,
"step": 458
},
{
"epoch": 2.466800804828974,
"grad_norm": 0.3246398479701492,
"learning_rate": 4.167152022937124e-06,
"loss": 1.1795,
"step": 459
},
{
"epoch": 2.4721663313212607,
"grad_norm": 0.33392434019026146,
"learning_rate": 4.085752861683981e-06,
"loss": 1.1344,
"step": 460
},
{
"epoch": 2.477531857813548,
"grad_norm": 0.32290812833166643,
"learning_rate": 4.005085768849856e-06,
"loss": 1.1691,
"step": 461
},
{
"epoch": 2.482897384305835,
"grad_norm": 0.3351539620126592,
"learning_rate": 3.925153568052123e-06,
"loss": 1.1926,
"step": 462
},
{
"epoch": 2.488262910798122,
"grad_norm": 0.32575534244471027,
"learning_rate": 3.845959057184453e-06,
"loss": 1.185,
"step": 463
},
{
"epoch": 2.493628437290409,
"grad_norm": 0.3318857239335513,
"learning_rate": 3.767505008318914e-06,
"loss": 1.1056,
"step": 464
},
{
"epoch": 2.4989939637826963,
"grad_norm": 0.3332508682060839,
"learning_rate": 3.689794167608937e-06,
"loss": 1.0975,
"step": 465
},
{
"epoch": 2.5043594902749833,
"grad_norm": 0.3322015968073437,
"learning_rate": 3.6128292551931924e-06,
"loss": 1.1779,
"step": 466
},
{
"epoch": 2.5097250167672702,
"grad_norm": 0.3339421172566103,
"learning_rate": 3.536612965100361e-06,
"loss": 1.1456,
"step": 467
},
{
"epoch": 2.515090543259557,
"grad_norm": 0.32193009926200084,
"learning_rate": 3.4611479651548457e-06,
"loss": 1.1247,
"step": 468
},
{
"epoch": 2.5204560697518446,
"grad_norm": 0.3275332021298932,
"learning_rate": 3.3864368968834074e-06,
"loss": 1.1583,
"step": 469
},
{
"epoch": 2.5258215962441315,
"grad_norm": 0.3322580175015273,
"learning_rate": 3.3124823754226624e-06,
"loss": 1.1773,
"step": 470
},
{
"epoch": 2.5258215962441315,
"eval_loss": 2.2257003784179688,
"eval_runtime": 403.3972,
"eval_samples_per_second": 6.319,
"eval_steps_per_second": 0.791,
"step": 470
},
{
"epoch": 2.5311871227364184,
"grad_norm": 0.3230969965574369,
"learning_rate": 3.239286989427573e-06,
"loss": 1.1593,
"step": 471
},
{
"epoch": 2.536552649228706,
"grad_norm": 0.3206756438379543,
"learning_rate": 3.166853300980821e-06,
"loss": 1.1583,
"step": 472
},
{
"epoch": 2.5419181757209928,
"grad_norm": 0.32211926813307257,
"learning_rate": 3.095183845503144e-06,
"loss": 1.1679,
"step": 473
},
{
"epoch": 2.5472837022132797,
"grad_norm": 0.326559743052197,
"learning_rate": 3.024281131664569e-06,
"loss": 1.1188,
"step": 474
},
{
"epoch": 2.5526492287055667,
"grad_norm": 0.3165584436088936,
"learning_rate": 2.9541476412966036e-06,
"loss": 1.1372,
"step": 475
},
{
"epoch": 2.5580147551978536,
"grad_norm": 0.31339823548327456,
"learning_rate": 2.884785829305381e-06,
"loss": 1.1235,
"step": 476
},
{
"epoch": 2.563380281690141,
"grad_norm": 0.33296469052198463,
"learning_rate": 2.8161981235857143e-06,
"loss": 1.0565,
"step": 477
},
{
"epoch": 2.568745808182428,
"grad_norm": 0.3222335104308849,
"learning_rate": 2.7483869249360915e-06,
"loss": 1.1068,
"step": 478
},
{
"epoch": 2.574111334674715,
"grad_norm": 0.3123105201965395,
"learning_rate": 2.681354606974698e-06,
"loss": 1.1675,
"step": 479
},
{
"epoch": 2.5794768611670023,
"grad_norm": 0.33382708976050407,
"learning_rate": 2.615103516056275e-06,
"loss": 1.1605,
"step": 480
},
{
"epoch": 2.584842387659289,
"grad_norm": 0.3218837631971954,
"learning_rate": 2.5496359711900118e-06,
"loss": 1.1689,
"step": 481
},
{
"epoch": 2.590207914151576,
"grad_norm": 0.31291467673725704,
"learning_rate": 2.484954263958383e-06,
"loss": 1.1812,
"step": 482
},
{
"epoch": 2.595573440643863,
"grad_norm": 0.32492352260184354,
"learning_rate": 2.4210606584369104e-06,
"loss": 1.1555,
"step": 483
},
{
"epoch": 2.60093896713615,
"grad_norm": 0.3504213880385304,
"learning_rate": 2.35795739111494e-06,
"loss": 1.1061,
"step": 484
},
{
"epoch": 2.6063044936284374,
"grad_norm": 0.31100084169291436,
"learning_rate": 2.2956466708173307e-06,
"loss": 1.0928,
"step": 485
},
{
"epoch": 2.6116700201207244,
"grad_norm": 0.3248255726906159,
"learning_rate": 2.2341306786271695e-06,
"loss": 1.1183,
"step": 486
},
{
"epoch": 2.6170355466130113,
"grad_norm": 0.31866373634955536,
"learning_rate": 2.173411567809394e-06,
"loss": 1.1753,
"step": 487
},
{
"epoch": 2.6224010731052987,
"grad_norm": 0.3352823580996406,
"learning_rate": 2.113491463735437e-06,
"loss": 1.1207,
"step": 488
},
{
"epoch": 2.6277665995975856,
"grad_norm": 0.31359112976859854,
"learning_rate": 2.0543724638088347e-06,
"loss": 1.1801,
"step": 489
},
{
"epoch": 2.6331321260898726,
"grad_norm": 0.33926000212154844,
"learning_rate": 1.996056637391805e-06,
"loss": 1.2241,
"step": 490
},
{
"epoch": 2.6384976525821595,
"grad_norm": 0.3371162704533217,
"learning_rate": 1.938546025732807e-06,
"loss": 1.0681,
"step": 491
},
{
"epoch": 2.6438631790744465,
"grad_norm": 0.3190932843505867,
"learning_rate": 1.881842641895104e-06,
"loss": 1.1484,
"step": 492
},
{
"epoch": 2.649228705566734,
"grad_norm": 0.3275133328320819,
"learning_rate": 1.8259484706862951e-06,
"loss": 1.1495,
"step": 493
},
{
"epoch": 2.654594232059021,
"grad_norm": 0.32869987839977655,
"learning_rate": 1.7708654685888337e-06,
"loss": 1.175,
"step": 494
},
{
"epoch": 2.6599597585513077,
"grad_norm": 0.3185931737498036,
"learning_rate": 1.7165955636915392e-06,
"loss": 1.071,
"step": 495
},
{
"epoch": 2.665325285043595,
"grad_norm": 0.31601212994979094,
"learning_rate": 1.6631406556221334e-06,
"loss": 1.1135,
"step": 496
},
{
"epoch": 2.670690811535882,
"grad_norm": 0.3252786042303215,
"learning_rate": 1.6105026154807218e-06,
"loss": 1.0958,
"step": 497
},
{
"epoch": 2.676056338028169,
"grad_norm": 0.3613550247333106,
"learning_rate": 1.558683285774304e-06,
"loss": 1.1663,
"step": 498
},
{
"epoch": 2.681421864520456,
"grad_norm": 0.3343048936672064,
"learning_rate": 1.5076844803522922e-06,
"loss": 1.2131,
"step": 499
},
{
"epoch": 2.686787391012743,
"grad_norm": 0.31812219422372123,
"learning_rate": 1.457507984343001e-06,
"loss": 1.2014,
"step": 500
},
{
"epoch": 2.6921529175050303,
"grad_norm": 0.31499628329755797,
"learning_rate": 1.408155554091184e-06,
"loss": 1.1767,
"step": 501
},
{
"epoch": 2.697518443997317,
"grad_norm": 0.3218866465290174,
"learning_rate": 1.3596289170965309e-06,
"loss": 1.1169,
"step": 502
},
{
"epoch": 2.702883970489604,
"grad_norm": 0.32278919495138014,
"learning_rate": 1.3119297719532242e-06,
"loss": 1.1428,
"step": 503
},
{
"epoch": 2.7082494969818915,
"grad_norm": 0.3249191581938583,
"learning_rate": 1.265059788290468e-06,
"loss": 1.1663,
"step": 504
},
{
"epoch": 2.7136150234741785,
"grad_norm": 0.3252541633067707,
"learning_rate": 1.2190206067140442e-06,
"loss": 1.2813,
"step": 505
},
{
"epoch": 2.7189805499664654,
"grad_norm": 0.3106253606000537,
"learning_rate": 1.173813838748894e-06,
"loss": 1.1415,
"step": 506
},
{
"epoch": 2.7243460764587524,
"grad_norm": 0.314646801186305,
"learning_rate": 1.129441066782702e-06,
"loss": 1.1541,
"step": 507
},
{
"epoch": 2.7297116029510393,
"grad_norm": 0.3213537454716302,
"learning_rate": 1.0859038440105164e-06,
"loss": 1.138,
"step": 508
},
{
"epoch": 2.7350771294433267,
"grad_norm": 0.32136955676763673,
"learning_rate": 1.0432036943803708e-06,
"loss": 1.1231,
"step": 509
},
{
"epoch": 2.7404426559356136,
"grad_norm": 0.3243563096847065,
"learning_rate": 1.0013421125399519e-06,
"loss": 1.244,
"step": 510
},
{
"epoch": 2.7458081824279006,
"grad_norm": 0.3089576646257491,
"learning_rate": 9.6032056378427e-07,
"loss": 1.1554,
"step": 511
},
{
"epoch": 2.751173708920188,
"grad_norm": 0.3095090303092117,
"learning_rate": 9.201404840043726e-07,
"loss": 1.1231,
"step": 512
},
{
"epoch": 2.756539235412475,
"grad_norm": 0.3147170138443251,
"learning_rate": 8.808032796371019e-07,
"loss": 1.0902,
"step": 513
},
{
"epoch": 2.761904761904762,
"grad_norm": 0.29507123132031615,
"learning_rate": 8.423103276158306e-07,
"loss": 1.0369,
"step": 514
},
{
"epoch": 2.767270288397049,
"grad_norm": 0.30105732124479273,
"learning_rate": 8.046629753222957e-07,
"loss": 1.1041,
"step": 515
},
{
"epoch": 2.7726358148893357,
"grad_norm": 0.3224663865586402,
"learning_rate": 7.678625405394157e-07,
"loss": 1.1015,
"step": 516
},
{
"epoch": 2.778001341381623,
"grad_norm": 0.3222742299894699,
"learning_rate": 7.319103114051707e-07,
"loss": 1.1864,
"step": 517
},
{
"epoch": 2.778001341381623,
"eval_loss": 2.225644826889038,
"eval_runtime": 403.2371,
"eval_samples_per_second": 6.321,
"eval_steps_per_second": 0.791,
"step": 517
},
{
"epoch": 2.78336686787391,
"grad_norm": 0.3171003364088583,
"learning_rate": 6.9680754636752e-07,
"loss": 1.1892,
"step": 518
},
{
"epoch": 2.788732394366197,
"grad_norm": 0.3119059918714888,
"learning_rate": 6.625554741403333e-07,
"loss": 1.2042,
"step": 519
},
{
"epoch": 2.7940979208584844,
"grad_norm": 0.3203009443769252,
"learning_rate": 6.291552936604134e-07,
"loss": 1.1335,
"step": 520
},
{
"epoch": 2.7994634473507714,
"grad_norm": 0.3288968824749739,
"learning_rate": 5.966081740454932e-07,
"loss": 1.192,
"step": 521
},
{
"epoch": 2.8048289738430583,
"grad_norm": 0.31460862008525864,
"learning_rate": 5.649152545533332e-07,
"loss": 1.1729,
"step": 522
},
{
"epoch": 2.8101945003353457,
"grad_norm": 0.3217212545480734,
"learning_rate": 5.340776445418472e-07,
"loss": 1.1193,
"step": 523
},
{
"epoch": 2.8155600268276326,
"grad_norm": 0.3123123100906072,
"learning_rate": 5.040964234302559e-07,
"loss": 1.133,
"step": 524
},
{
"epoch": 2.8209255533199196,
"grad_norm": 0.31788975020445803,
"learning_rate": 4.749726406613142e-07,
"loss": 1.1569,
"step": 525
},
{
"epoch": 2.8262910798122065,
"grad_norm": 0.35427580228034644,
"learning_rate": 4.4670731566457126e-07,
"loss": 1.185,
"step": 526
},
{
"epoch": 2.8316566063044935,
"grad_norm": 0.3078993071472304,
"learning_rate": 4.1930143782070444e-07,
"loss": 1.245,
"step": 527
},
{
"epoch": 2.837022132796781,
"grad_norm": 0.627826963441227,
"learning_rate": 3.9275596642685543e-07,
"loss": 1.1768,
"step": 528
},
{
"epoch": 2.842387659289068,
"grad_norm": 0.3180386319315376,
"learning_rate": 3.670718306630766e-07,
"loss": 1.1915,
"step": 529
},
{
"epoch": 2.8477531857813547,
"grad_norm": 0.3183014815220175,
"learning_rate": 3.422499295598069e-07,
"loss": 1.1618,
"step": 530
},
{
"epoch": 2.853118712273642,
"grad_norm": 0.31162900762652046,
"learning_rate": 3.1829113196638614e-07,
"loss": 1.1459,
"step": 531
},
{
"epoch": 2.858484238765929,
"grad_norm": 0.31454238955052416,
"learning_rate": 2.9519627652065674e-07,
"loss": 1.1396,
"step": 532
},
{
"epoch": 2.863849765258216,
"grad_norm": 0.33536803393906284,
"learning_rate": 2.7296617161960415e-07,
"loss": 1.1325,
"step": 533
},
{
"epoch": 2.869215291750503,
"grad_norm": 0.3118341382685939,
"learning_rate": 2.5160159539105443e-07,
"loss": 1.186,
"step": 534
},
{
"epoch": 2.87458081824279,
"grad_norm": 0.30550007371594207,
"learning_rate": 2.3110329566645162e-07,
"loss": 1.1615,
"step": 535
},
{
"epoch": 2.8799463447350773,
"grad_norm": 0.3290186323857588,
"learning_rate": 2.1147198995466467e-07,
"loss": 1.1199,
"step": 536
},
{
"epoch": 2.885311871227364,
"grad_norm": 0.30975536579912843,
"learning_rate": 1.927083654168854e-07,
"loss": 1.1207,
"step": 537
},
{
"epoch": 2.890677397719651,
"grad_norm": 0.3272675259115993,
"learning_rate": 1.7481307884256727e-07,
"loss": 1.1211,
"step": 538
},
{
"epoch": 2.8960429242119385,
"grad_norm": 0.3186123525466991,
"learning_rate": 1.5778675662643793e-07,
"loss": 1.2159,
"step": 539
},
{
"epoch": 2.9014084507042255,
"grad_norm": 0.3238665460415356,
"learning_rate": 1.4162999474657268e-07,
"loss": 1.1226,
"step": 540
},
{
"epoch": 2.9067739771965124,
"grad_norm": 0.300927276396165,
"learning_rate": 1.2634335874353585e-07,
"loss": 1.1584,
"step": 541
},
{
"epoch": 2.9121395036887994,
"grad_norm": 0.3225080409751762,
"learning_rate": 1.1192738370058575e-07,
"loss": 1.1549,
"step": 542
},
{
"epoch": 2.9175050301810863,
"grad_norm": 0.3014139103111567,
"learning_rate": 9.838257422493668e-08,
"loss": 1.0577,
"step": 543
},
{
"epoch": 2.9228705566733737,
"grad_norm": 0.33343711969938683,
"learning_rate": 8.570940443010655e-08,
"loss": 1.1204,
"step": 544
},
{
"epoch": 2.9282360831656606,
"grad_norm": 0.3001682609568673,
"learning_rate": 7.390831791931896e-08,
"loss": 1.116,
"step": 545
},
{
"epoch": 2.9336016096579476,
"grad_norm": 0.3132232444134447,
"learning_rate": 6.297972776996286e-08,
"loss": 1.1668,
"step": 546
},
{
"epoch": 2.938967136150235,
"grad_norm": 0.317135513260303,
"learning_rate": 5.2924016519156906e-08,
"loss": 1.1322,
"step": 547
},
{
"epoch": 2.944332662642522,
"grad_norm": 0.324986722470258,
"learning_rate": 4.374153615033794e-08,
"loss": 1.1091,
"step": 548
},
{
"epoch": 2.949698189134809,
"grad_norm": 0.30922413089636414,
"learning_rate": 3.543260808095139e-08,
"loss": 1.141,
"step": 549
},
{
"epoch": 2.955063715627096,
"grad_norm": 0.31855167384942973,
"learning_rate": 2.799752315119919e-08,
"loss": 1.2269,
"step": 550
},
{
"epoch": 2.9604292421193827,
"grad_norm": 0.3074014352539391,
"learning_rate": 2.1436541613853444e-08,
"loss": 1.1442,
"step": 551
},
{
"epoch": 2.96579476861167,
"grad_norm": 0.3064677206528894,
"learning_rate": 1.5749893125160954e-08,
"loss": 1.1533,
"step": 552
},
{
"epoch": 2.971160295103957,
"grad_norm": 0.3049690197063334,
"learning_rate": 1.0937776736782978e-08,
"loss": 1.112,
"step": 553
},
{
"epoch": 2.976525821596244,
"grad_norm": 0.3030005645927827,
"learning_rate": 7.00036088885081e-09,
"loss": 1.1847,
"step": 554
},
{
"epoch": 2.9818913480885314,
"grad_norm": 0.32049007711358624,
"learning_rate": 3.9377834040538185e-09,
"loss": 1.1018,
"step": 555
},
{
"epoch": 2.9872568745808183,
"grad_norm": 0.3209495594839988,
"learning_rate": 1.7501514828183185e-09,
"loss": 1.1483,
"step": 556
},
{
"epoch": 2.9926224010731053,
"grad_norm": 0.3215801377358293,
"learning_rate": 4.3754169955778634e-10,
"loss": 1.1563,
"step": 557
},
{
"epoch": 2.9979879275653922,
"grad_norm": 0.29931532757386853,
"learning_rate": 0.0,
"loss": 1.1306,
"step": 558
}
],
"logging_steps": 1,
"max_steps": 558,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 47,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.4979627422066934e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}