nemo-normal-e2.0 / trainer_state.json
kalomaze's picture
Upload folder using huggingface_hub
7142bf4 verified
raw
history blame
141 kB
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 500,
"global_step": 814,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.002457002457002457,
"grad_norm": 12.175590846877855,
"learning_rate": 0.0,
"loss": 1.2148,
"step": 1
},
{
"epoch": 0.004914004914004914,
"grad_norm": 13.112846331209713,
"learning_rate": 5e-08,
"loss": 1.3154,
"step": 2
},
{
"epoch": 0.007371007371007371,
"grad_norm": 12.768616794783732,
"learning_rate": 1e-07,
"loss": 1.2865,
"step": 3
},
{
"epoch": 0.009828009828009828,
"grad_norm": 12.570106090577944,
"learning_rate": 1.5e-07,
"loss": 1.2459,
"step": 4
},
{
"epoch": 0.012285012285012284,
"grad_norm": 12.026844749086528,
"learning_rate": 2e-07,
"loss": 1.2496,
"step": 5
},
{
"epoch": 0.014742014742014743,
"grad_norm": 12.166619597781125,
"learning_rate": 2.5e-07,
"loss": 1.2179,
"step": 6
},
{
"epoch": 0.0171990171990172,
"grad_norm": 12.450604071787023,
"learning_rate": 3e-07,
"loss": 1.2451,
"step": 7
},
{
"epoch": 0.019656019656019656,
"grad_norm": 11.75536378463392,
"learning_rate": 3.5e-07,
"loss": 1.2372,
"step": 8
},
{
"epoch": 0.022113022113022112,
"grad_norm": 11.7494451205682,
"learning_rate": 4e-07,
"loss": 1.2174,
"step": 9
},
{
"epoch": 0.02457002457002457,
"grad_norm": 11.485164591676915,
"learning_rate": 4.5e-07,
"loss": 1.1803,
"step": 10
},
{
"epoch": 0.02702702702702703,
"grad_norm": 11.41671410777608,
"learning_rate": 5e-07,
"loss": 1.1829,
"step": 11
},
{
"epoch": 0.029484029484029485,
"grad_norm": 10.737969644794076,
"learning_rate": 5.5e-07,
"loss": 1.1912,
"step": 12
},
{
"epoch": 0.03194103194103194,
"grad_norm": 12.0315769410197,
"learning_rate": 6e-07,
"loss": 1.2516,
"step": 13
},
{
"epoch": 0.0343980343980344,
"grad_norm": 11.703183966857443,
"learning_rate": 6.5e-07,
"loss": 1.2533,
"step": 14
},
{
"epoch": 0.036855036855036855,
"grad_norm": 10.75777994461547,
"learning_rate": 7e-07,
"loss": 1.1763,
"step": 15
},
{
"epoch": 0.03931203931203931,
"grad_norm": 9.517567846059583,
"learning_rate": 7.5e-07,
"loss": 1.1474,
"step": 16
},
{
"epoch": 0.04176904176904177,
"grad_norm": 9.825919996727293,
"learning_rate": 8e-07,
"loss": 1.1725,
"step": 17
},
{
"epoch": 0.044226044226044224,
"grad_norm": 9.370499752545927,
"learning_rate": 8.499999999999999e-07,
"loss": 1.2048,
"step": 18
},
{
"epoch": 0.04668304668304668,
"grad_norm": 8.582608952468464,
"learning_rate": 9e-07,
"loss": 1.1428,
"step": 19
},
{
"epoch": 0.04914004914004914,
"grad_norm": 7.4140901371154,
"learning_rate": 9.499999999999999e-07,
"loss": 1.1583,
"step": 20
},
{
"epoch": 0.051597051597051594,
"grad_norm": 6.890519345341131,
"learning_rate": 1e-06,
"loss": 1.1961,
"step": 21
},
{
"epoch": 0.05405405405405406,
"grad_norm": 5.47854198998431,
"learning_rate": 1.05e-06,
"loss": 1.126,
"step": 22
},
{
"epoch": 0.056511056511056514,
"grad_norm": 5.350914917259245,
"learning_rate": 1.1e-06,
"loss": 1.1484,
"step": 23
},
{
"epoch": 0.05896805896805897,
"grad_norm": 4.114918765551221,
"learning_rate": 1.1499999999999998e-06,
"loss": 1.1328,
"step": 24
},
{
"epoch": 0.06142506142506143,
"grad_norm": 3.535376327350945,
"learning_rate": 1.2e-06,
"loss": 1.0889,
"step": 25
},
{
"epoch": 0.06388206388206388,
"grad_norm": 3.2480152784573586,
"learning_rate": 1.2499999999999999e-06,
"loss": 1.0437,
"step": 26
},
{
"epoch": 0.06633906633906633,
"grad_norm": 3.4795119249436293,
"learning_rate": 1.3e-06,
"loss": 1.0966,
"step": 27
},
{
"epoch": 0.0687960687960688,
"grad_norm": 2.8566181371095047,
"learning_rate": 1.35e-06,
"loss": 1.1234,
"step": 28
},
{
"epoch": 0.07125307125307126,
"grad_norm": 3.6578709990785048,
"learning_rate": 1.4e-06,
"loss": 1.1265,
"step": 29
},
{
"epoch": 0.07371007371007371,
"grad_norm": 3.541143081282278,
"learning_rate": 1.4499999999999999e-06,
"loss": 1.0543,
"step": 30
},
{
"epoch": 0.07616707616707617,
"grad_norm": 6.531880288467122,
"learning_rate": 1.5e-06,
"loss": 1.1109,
"step": 31
},
{
"epoch": 0.07862407862407862,
"grad_norm": 6.4715106690759,
"learning_rate": 1.55e-06,
"loss": 1.1257,
"step": 32
},
{
"epoch": 0.08108108108108109,
"grad_norm": 5.399614723171929,
"learning_rate": 1.6e-06,
"loss": 1.0694,
"step": 33
},
{
"epoch": 0.08353808353808354,
"grad_norm": 5.447520160063898,
"learning_rate": 1.6499999999999999e-06,
"loss": 1.047,
"step": 34
},
{
"epoch": 0.085995085995086,
"grad_norm": 15.759399235218536,
"learning_rate": 1.6999999999999998e-06,
"loss": 1.0684,
"step": 35
},
{
"epoch": 0.08845208845208845,
"grad_norm": 6.138108825882396,
"learning_rate": 1.75e-06,
"loss": 1.0568,
"step": 36
},
{
"epoch": 0.09090909090909091,
"grad_norm": 28.96303499581736,
"learning_rate": 1.8e-06,
"loss": 1.0748,
"step": 37
},
{
"epoch": 0.09336609336609336,
"grad_norm": 3.400779377353331,
"learning_rate": 1.85e-06,
"loss": 1.0841,
"step": 38
},
{
"epoch": 0.09582309582309582,
"grad_norm": 3.2088443423156305,
"learning_rate": 1.8999999999999998e-06,
"loss": 1.0637,
"step": 39
},
{
"epoch": 0.09828009828009827,
"grad_norm": 3.650839591876822,
"learning_rate": 1.95e-06,
"loss": 1.065,
"step": 40
},
{
"epoch": 0.10073710073710074,
"grad_norm": 2.48874031529166,
"learning_rate": 2e-06,
"loss": 1.0325,
"step": 41
},
{
"epoch": 0.10319410319410319,
"grad_norm": 2.4764261500155436,
"learning_rate": 1.9999917626554467e-06,
"loss": 1.0484,
"step": 42
},
{
"epoch": 0.10565110565110565,
"grad_norm": 2.0092623603336794,
"learning_rate": 1.9999670507574944e-06,
"loss": 1.0859,
"step": 43
},
{
"epoch": 0.10810810810810811,
"grad_norm": 1.915125996872455,
"learning_rate": 1.9999258647132643e-06,
"loss": 1.0423,
"step": 44
},
{
"epoch": 0.11056511056511056,
"grad_norm": 1.8459238588876408,
"learning_rate": 1.9998682052012837e-06,
"loss": 1.0686,
"step": 45
},
{
"epoch": 0.11302211302211303,
"grad_norm": 2.238673168744122,
"learning_rate": 1.9997940731714745e-06,
"loss": 1.0601,
"step": 46
},
{
"epoch": 0.11547911547911548,
"grad_norm": 4.306515287574078,
"learning_rate": 1.9997034698451393e-06,
"loss": 1.0367,
"step": 47
},
{
"epoch": 0.11793611793611794,
"grad_norm": 5.689973830908464,
"learning_rate": 1.9995963967149398e-06,
"loss": 1.0542,
"step": 48
},
{
"epoch": 0.12039312039312039,
"grad_norm": 2.1880376147294522,
"learning_rate": 1.999472855544872e-06,
"loss": 1.0635,
"step": 49
},
{
"epoch": 0.12285012285012285,
"grad_norm": 3.433335591384566,
"learning_rate": 1.999332848370239e-06,
"loss": 1.0323,
"step": 50
},
{
"epoch": 0.12530712530712532,
"grad_norm": 1.8169785229927564,
"learning_rate": 1.9991763774976158e-06,
"loss": 1.0218,
"step": 51
},
{
"epoch": 0.12776412776412777,
"grad_norm": 1.8670733937364128,
"learning_rate": 1.9990034455048096e-06,
"loss": 1.0405,
"step": 52
},
{
"epoch": 0.13022113022113022,
"grad_norm": 2.439448369982022,
"learning_rate": 1.998814055240823e-06,
"loss": 1.0519,
"step": 53
},
{
"epoch": 0.13267813267813267,
"grad_norm": 1.5940089652939573,
"learning_rate": 1.9986082098258006e-06,
"loss": 1.0359,
"step": 54
},
{
"epoch": 0.13513513513513514,
"grad_norm": 1.7404888899145354,
"learning_rate": 1.9983859126509824e-06,
"loss": 1.0087,
"step": 55
},
{
"epoch": 0.1375921375921376,
"grad_norm": 2.2476917865688026,
"learning_rate": 1.998147167378645e-06,
"loss": 1.0138,
"step": 56
},
{
"epoch": 0.14004914004914004,
"grad_norm": 1.8002247115537675,
"learning_rate": 1.9978919779420425e-06,
"loss": 1.0488,
"step": 57
},
{
"epoch": 0.14250614250614252,
"grad_norm": 1.663251752279236,
"learning_rate": 1.9976203485453413e-06,
"loss": 0.9809,
"step": 58
},
{
"epoch": 0.14496314496314497,
"grad_norm": 1.7653797682084729,
"learning_rate": 1.9973322836635515e-06,
"loss": 1.0686,
"step": 59
},
{
"epoch": 0.14742014742014742,
"grad_norm": 1.5336222427197481,
"learning_rate": 1.9970277880424526e-06,
"loss": 0.986,
"step": 60
},
{
"epoch": 0.14987714987714987,
"grad_norm": 1.4421704567479543,
"learning_rate": 1.9967068666985146e-06,
"loss": 0.9948,
"step": 61
},
{
"epoch": 0.15233415233415235,
"grad_norm": 1.4042358001427657,
"learning_rate": 1.996369524918818e-06,
"loss": 0.967,
"step": 62
},
{
"epoch": 0.1547911547911548,
"grad_norm": 1.4295825578558987,
"learning_rate": 1.9960157682609633e-06,
"loss": 0.9723,
"step": 63
},
{
"epoch": 0.15724815724815724,
"grad_norm": 1.6093180146467703,
"learning_rate": 1.9956456025529804e-06,
"loss": 0.9737,
"step": 64
},
{
"epoch": 0.1597051597051597,
"grad_norm": 1.7273275051707897,
"learning_rate": 1.9952590338932358e-06,
"loss": 0.9935,
"step": 65
},
{
"epoch": 0.16216216216216217,
"grad_norm": 1.647950655242118,
"learning_rate": 1.994856068650327e-06,
"loss": 1.0509,
"step": 66
},
{
"epoch": 0.16461916461916462,
"grad_norm": 1.6052330571582885,
"learning_rate": 1.994436713462982e-06,
"loss": 1.0295,
"step": 67
},
{
"epoch": 0.16707616707616707,
"grad_norm": 1.694905424784479,
"learning_rate": 1.994000975239946e-06,
"loss": 1.0256,
"step": 68
},
{
"epoch": 0.16953316953316952,
"grad_norm": 1.5039835360361733,
"learning_rate": 1.9935488611598713e-06,
"loss": 1.0347,
"step": 69
},
{
"epoch": 0.171990171990172,
"grad_norm": 1.734126157202943,
"learning_rate": 1.9930803786711967e-06,
"loss": 1.023,
"step": 70
},
{
"epoch": 0.17444717444717445,
"grad_norm": 1.5242931763881968,
"learning_rate": 1.9925955354920264e-06,
"loss": 1.0406,
"step": 71
},
{
"epoch": 0.1769041769041769,
"grad_norm": 1.606370733528827,
"learning_rate": 1.99209433961e-06,
"loss": 1.0222,
"step": 72
},
{
"epoch": 0.17936117936117937,
"grad_norm": 2.4762050432839318,
"learning_rate": 1.991576799282164e-06,
"loss": 1.0278,
"step": 73
},
{
"epoch": 0.18181818181818182,
"grad_norm": 1.4747727870105878,
"learning_rate": 1.9910429230348343e-06,
"loss": 0.9779,
"step": 74
},
{
"epoch": 0.18427518427518427,
"grad_norm": 2.0977776864765167,
"learning_rate": 1.990492719663457e-06,
"loss": 1.0235,
"step": 75
},
{
"epoch": 0.18673218673218672,
"grad_norm": 1.9728896794221442,
"learning_rate": 1.9899261982324607e-06,
"loss": 0.9926,
"step": 76
},
{
"epoch": 0.1891891891891892,
"grad_norm": 1.7736978075188574,
"learning_rate": 1.98934336807511e-06,
"loss": 1.0354,
"step": 77
},
{
"epoch": 0.19164619164619165,
"grad_norm": 2.1716440393079464,
"learning_rate": 1.988744238793351e-06,
"loss": 0.977,
"step": 78
},
{
"epoch": 0.1941031941031941,
"grad_norm": 1.599345820298087,
"learning_rate": 1.9881288202576518e-06,
"loss": 0.9685,
"step": 79
},
{
"epoch": 0.19656019656019655,
"grad_norm": 1.4582630785882413,
"learning_rate": 1.9874971226068415e-06,
"loss": 0.943,
"step": 80
},
{
"epoch": 0.19901719901719903,
"grad_norm": 1.4275025088638074,
"learning_rate": 1.9868491562479424e-06,
"loss": 1.0295,
"step": 81
},
{
"epoch": 0.20147420147420148,
"grad_norm": 1.609247365977497,
"learning_rate": 1.9861849318559994e-06,
"loss": 0.9531,
"step": 82
},
{
"epoch": 0.20393120393120392,
"grad_norm": 1.711398906589491,
"learning_rate": 1.9855044603739028e-06,
"loss": 1.0302,
"step": 83
},
{
"epoch": 0.20638820638820637,
"grad_norm": 1.56099328882278,
"learning_rate": 1.984807753012208e-06,
"loss": 0.9684,
"step": 84
},
{
"epoch": 0.20884520884520885,
"grad_norm": 1.5428758027146767,
"learning_rate": 1.9840948212489527e-06,
"loss": 1.0129,
"step": 85
},
{
"epoch": 0.2113022113022113,
"grad_norm": 1.9669978126079752,
"learning_rate": 1.983365676829466e-06,
"loss": 0.9976,
"step": 86
},
{
"epoch": 0.21375921375921375,
"grad_norm": 1.5144216325008544,
"learning_rate": 1.9826203317661756e-06,
"loss": 0.9967,
"step": 87
},
{
"epoch": 0.21621621621621623,
"grad_norm": 2.1395022590018367,
"learning_rate": 1.9818587983384094e-06,
"loss": 0.9693,
"step": 88
},
{
"epoch": 0.21867321867321868,
"grad_norm": 1.5091268331439933,
"learning_rate": 1.981081089092194e-06,
"loss": 0.9271,
"step": 89
},
{
"epoch": 0.22113022113022113,
"grad_norm": 1.8173026339961775,
"learning_rate": 1.9802872168400476e-06,
"loss": 1.0819,
"step": 90
},
{
"epoch": 0.22358722358722358,
"grad_norm": 1.5986438761107704,
"learning_rate": 1.979477194660769e-06,
"loss": 0.9594,
"step": 91
},
{
"epoch": 0.22604422604422605,
"grad_norm": 1.5027188037219825,
"learning_rate": 1.978651035899221e-06,
"loss": 0.9964,
"step": 92
},
{
"epoch": 0.2285012285012285,
"grad_norm": 1.3856569445993616,
"learning_rate": 1.977808754166113e-06,
"loss": 0.9476,
"step": 93
},
{
"epoch": 0.23095823095823095,
"grad_norm": 1.445303089718021,
"learning_rate": 1.976950363337774e-06,
"loss": 0.9664,
"step": 94
},
{
"epoch": 0.2334152334152334,
"grad_norm": 1.9267867671090184,
"learning_rate": 1.9760758775559273e-06,
"loss": 1.0446,
"step": 95
},
{
"epoch": 0.23587223587223588,
"grad_norm": 1.4687411243380588,
"learning_rate": 1.975185311227453e-06,
"loss": 0.9483,
"step": 96
},
{
"epoch": 0.23832923832923833,
"grad_norm": 1.5489115105120785,
"learning_rate": 1.9742786790241546e-06,
"loss": 1.0073,
"step": 97
},
{
"epoch": 0.24078624078624078,
"grad_norm": 2.5463794252310676,
"learning_rate": 1.9733559958825166e-06,
"loss": 1.0046,
"step": 98
},
{
"epoch": 0.24324324324324326,
"grad_norm": 1.56125807056489,
"learning_rate": 1.9724172770034566e-06,
"loss": 0.971,
"step": 99
},
{
"epoch": 0.2457002457002457,
"grad_norm": 1.740701138818393,
"learning_rate": 1.971462537852076e-06,
"loss": 0.9856,
"step": 100
},
{
"epoch": 0.24815724815724816,
"grad_norm": 1.4470402917114091,
"learning_rate": 1.970491794157405e-06,
"loss": 0.9897,
"step": 101
},
{
"epoch": 0.25061425061425063,
"grad_norm": 1.5706400471340238,
"learning_rate": 1.9695050619121456e-06,
"loss": 0.9328,
"step": 102
},
{
"epoch": 0.25307125307125306,
"grad_norm": 1.9773708156391945,
"learning_rate": 1.9685023573724034e-06,
"loss": 0.9719,
"step": 103
},
{
"epoch": 0.25552825552825553,
"grad_norm": 1.9195746558098163,
"learning_rate": 1.9674836970574253e-06,
"loss": 0.978,
"step": 104
},
{
"epoch": 0.257985257985258,
"grad_norm": 1.8027098570962006,
"learning_rate": 1.966449097749322e-06,
"loss": 1.0023,
"step": 105
},
{
"epoch": 0.26044226044226043,
"grad_norm": 1.7805523167150648,
"learning_rate": 1.965398576492796e-06,
"loss": 0.9768,
"step": 106
},
{
"epoch": 0.2628992628992629,
"grad_norm": 1.9751836869745634,
"learning_rate": 1.9643321505948585e-06,
"loss": 1.0132,
"step": 107
},
{
"epoch": 0.26535626535626533,
"grad_norm": 1.6135714113454362,
"learning_rate": 1.9632498376245444e-06,
"loss": 1.0169,
"step": 108
},
{
"epoch": 0.2678132678132678,
"grad_norm": 1.6366575829852479,
"learning_rate": 1.9621516554126237e-06,
"loss": 0.9618,
"step": 109
},
{
"epoch": 0.2702702702702703,
"grad_norm": 1.4439243574845881,
"learning_rate": 1.9610376220513063e-06,
"loss": 1.0111,
"step": 110
},
{
"epoch": 0.2727272727272727,
"grad_norm": 1.574054550140938,
"learning_rate": 1.9599077558939466e-06,
"loss": 0.9979,
"step": 111
},
{
"epoch": 0.2751842751842752,
"grad_norm": 1.9768484700868765,
"learning_rate": 1.958762075554737e-06,
"loss": 1.0266,
"step": 112
},
{
"epoch": 0.27764127764127766,
"grad_norm": 1.6821288693476542,
"learning_rate": 1.957600599908406e-06,
"loss": 0.983,
"step": 113
},
{
"epoch": 0.2800982800982801,
"grad_norm": 1.6250908693533597,
"learning_rate": 1.956423348089903e-06,
"loss": 0.9735,
"step": 114
},
{
"epoch": 0.28255528255528256,
"grad_norm": 1.5158186810772534,
"learning_rate": 1.9552303394940857e-06,
"loss": 0.9767,
"step": 115
},
{
"epoch": 0.28501228501228504,
"grad_norm": 1.7591914283817052,
"learning_rate": 1.9540215937754004e-06,
"loss": 0.9515,
"step": 116
},
{
"epoch": 0.28746928746928746,
"grad_norm": 1.329980995859771,
"learning_rate": 1.952797130847557e-06,
"loss": 0.944,
"step": 117
},
{
"epoch": 0.28992628992628994,
"grad_norm": 1.7122579873222803,
"learning_rate": 1.951556970883201e-06,
"loss": 1.0098,
"step": 118
},
{
"epoch": 0.29238329238329236,
"grad_norm": 1.8427163819779844,
"learning_rate": 1.9503011343135825e-06,
"loss": 0.9815,
"step": 119
},
{
"epoch": 0.29484029484029484,
"grad_norm": 1.3722808372435795,
"learning_rate": 1.9490296418282184e-06,
"loss": 0.9914,
"step": 120
},
{
"epoch": 0.2972972972972973,
"grad_norm": 1.2858110734745247,
"learning_rate": 1.9477425143745523e-06,
"loss": 0.9727,
"step": 121
},
{
"epoch": 0.29975429975429974,
"grad_norm": 1.3478569354201173,
"learning_rate": 1.9464397731576094e-06,
"loss": 0.9557,
"step": 122
},
{
"epoch": 0.3022113022113022,
"grad_norm": 1.534212309472206,
"learning_rate": 1.945121439639645e-06,
"loss": 0.9995,
"step": 123
},
{
"epoch": 0.3046683046683047,
"grad_norm": 1.4583668620611645,
"learning_rate": 1.943787535539795e-06,
"loss": 0.9244,
"step": 124
},
{
"epoch": 0.3071253071253071,
"grad_norm": 1.3048849492711914,
"learning_rate": 1.9424380828337143e-06,
"loss": 0.9749,
"step": 125
},
{
"epoch": 0.3095823095823096,
"grad_norm": 1.9214890498265684,
"learning_rate": 1.9410731037532167e-06,
"loss": 1.0126,
"step": 126
},
{
"epoch": 0.31203931203931207,
"grad_norm": 1.492998355230107,
"learning_rate": 1.9396926207859082e-06,
"loss": 0.9216,
"step": 127
},
{
"epoch": 0.3144963144963145,
"grad_norm": 1.529088663498555,
"learning_rate": 1.9382966566748166e-06,
"loss": 1.0266,
"step": 128
},
{
"epoch": 0.31695331695331697,
"grad_norm": 1.4961622214991601,
"learning_rate": 1.9368852344180167e-06,
"loss": 0.949,
"step": 129
},
{
"epoch": 0.3194103194103194,
"grad_norm": 1.5657840764329438,
"learning_rate": 1.935458377268251e-06,
"loss": 0.9768,
"step": 130
},
{
"epoch": 0.32186732186732187,
"grad_norm": 1.458088079896784,
"learning_rate": 1.934016108732548e-06,
"loss": 0.9831,
"step": 131
},
{
"epoch": 0.32432432432432434,
"grad_norm": 1.5704249898970843,
"learning_rate": 1.932558452571833e-06,
"loss": 0.9845,
"step": 132
},
{
"epoch": 0.32678132678132676,
"grad_norm": 1.366875012537187,
"learning_rate": 1.9310854328005377e-06,
"loss": 0.9751,
"step": 133
},
{
"epoch": 0.32923832923832924,
"grad_norm": 1.6682519282454646,
"learning_rate": 1.9295970736862064e-06,
"loss": 0.9278,
"step": 134
},
{
"epoch": 0.3316953316953317,
"grad_norm": 1.7010058864881312,
"learning_rate": 1.928093399749091e-06,
"loss": 0.9479,
"step": 135
},
{
"epoch": 0.33415233415233414,
"grad_norm": 2.774339563092048,
"learning_rate": 1.926574435761753e-06,
"loss": 0.9216,
"step": 136
},
{
"epoch": 0.3366093366093366,
"grad_norm": 1.5996469164653206,
"learning_rate": 1.925040206748652e-06,
"loss": 0.9749,
"step": 137
},
{
"epoch": 0.33906633906633904,
"grad_norm": 1.6140642370600065,
"learning_rate": 1.9234907379857334e-06,
"loss": 1.0195,
"step": 138
},
{
"epoch": 0.3415233415233415,
"grad_norm": 1.6724211149688903,
"learning_rate": 1.9219260550000143e-06,
"loss": 1.0086,
"step": 139
},
{
"epoch": 0.343980343980344,
"grad_norm": 1.6697054307200352,
"learning_rate": 1.920346183569159e-06,
"loss": 1.0177,
"step": 140
},
{
"epoch": 0.3464373464373464,
"grad_norm": 1.5507362171936665,
"learning_rate": 1.91875114972106e-06,
"loss": 0.9701,
"step": 141
},
{
"epoch": 0.3488943488943489,
"grad_norm": 1.572772556059398,
"learning_rate": 1.9171409797334026e-06,
"loss": 0.9893,
"step": 142
},
{
"epoch": 0.35135135135135137,
"grad_norm": 1.4851773686144312,
"learning_rate": 1.9155157001332373e-06,
"loss": 0.9617,
"step": 143
},
{
"epoch": 0.3538083538083538,
"grad_norm": 1.3935370189220326,
"learning_rate": 1.91387533769654e-06,
"loss": 0.9647,
"step": 144
},
{
"epoch": 0.35626535626535627,
"grad_norm": 2.086032038197472,
"learning_rate": 1.912219919447772e-06,
"loss": 1.0138,
"step": 145
},
{
"epoch": 0.35872235872235875,
"grad_norm": 1.8988521674993861,
"learning_rate": 1.9105494726594343e-06,
"loss": 0.9619,
"step": 146
},
{
"epoch": 0.36117936117936117,
"grad_norm": 1.7707337422933225,
"learning_rate": 1.9088640248516184e-06,
"loss": 0.9689,
"step": 147
},
{
"epoch": 0.36363636363636365,
"grad_norm": 10.58019032383696,
"learning_rate": 1.9071636037915533e-06,
"loss": 0.9886,
"step": 148
},
{
"epoch": 0.36609336609336607,
"grad_norm": 1.6326349036691947,
"learning_rate": 1.9054482374931466e-06,
"loss": 0.9928,
"step": 149
},
{
"epoch": 0.36855036855036855,
"grad_norm": 1.4602979723959129,
"learning_rate": 1.9037179542165251e-06,
"loss": 0.9451,
"step": 150
},
{
"epoch": 0.371007371007371,
"grad_norm": 1.5983113709927816,
"learning_rate": 1.9019727824675682e-06,
"loss": 0.9488,
"step": 151
},
{
"epoch": 0.37346437346437344,
"grad_norm": 1.394284340663908,
"learning_rate": 1.9002127509974373e-06,
"loss": 0.9717,
"step": 152
},
{
"epoch": 0.3759213759213759,
"grad_norm": 1.856689232868256,
"learning_rate": 1.8984378888021041e-06,
"loss": 0.9286,
"step": 153
},
{
"epoch": 0.3783783783783784,
"grad_norm": 1.5404154844189102,
"learning_rate": 1.8966482251218714e-06,
"loss": 0.9625,
"step": 154
},
{
"epoch": 0.3808353808353808,
"grad_norm": 1.5417688714414741,
"learning_rate": 1.8948437894408916e-06,
"loss": 0.9314,
"step": 155
},
{
"epoch": 0.3832923832923833,
"grad_norm": 1.8457706124864284,
"learning_rate": 1.8930246114866822e-06,
"loss": 1.0147,
"step": 156
},
{
"epoch": 0.3857493857493858,
"grad_norm": 1.4971057129868792,
"learning_rate": 1.8911907212296341e-06,
"loss": 0.9309,
"step": 157
},
{
"epoch": 0.3882063882063882,
"grad_norm": 1.6347670467700182,
"learning_rate": 1.8893421488825188e-06,
"loss": 0.9621,
"step": 158
},
{
"epoch": 0.3906633906633907,
"grad_norm": 1.6324227560172353,
"learning_rate": 1.8874789248999915e-06,
"loss": 0.9772,
"step": 159
},
{
"epoch": 0.3931203931203931,
"grad_norm": 1.6267750782874413,
"learning_rate": 1.885601079978088e-06,
"loss": 0.9861,
"step": 160
},
{
"epoch": 0.3955773955773956,
"grad_norm": 1.4406136327704948,
"learning_rate": 1.8837086450537192e-06,
"loss": 0.9406,
"step": 161
},
{
"epoch": 0.39803439803439805,
"grad_norm": 1.5622857080744104,
"learning_rate": 1.8818016513041624e-06,
"loss": 0.94,
"step": 162
},
{
"epoch": 0.4004914004914005,
"grad_norm": 1.7312279395333572,
"learning_rate": 1.8798801301465468e-06,
"loss": 0.96,
"step": 163
},
{
"epoch": 0.40294840294840295,
"grad_norm": 1.5736688729195003,
"learning_rate": 1.877944113237336e-06,
"loss": 0.9793,
"step": 164
},
{
"epoch": 0.40540540540540543,
"grad_norm": 1.501899237192728,
"learning_rate": 1.8759936324718066e-06,
"loss": 0.9178,
"step": 165
},
{
"epoch": 0.40786240786240785,
"grad_norm": 1.3483386139829852,
"learning_rate": 1.874028719983523e-06,
"loss": 0.9593,
"step": 166
},
{
"epoch": 0.4103194103194103,
"grad_norm": 1.5326296765065217,
"learning_rate": 1.8720494081438077e-06,
"loss": 0.9794,
"step": 167
},
{
"epoch": 0.41277641277641275,
"grad_norm": 1.5759774818522712,
"learning_rate": 1.870055729561207e-06,
"loss": 0.9418,
"step": 168
},
{
"epoch": 0.4152334152334152,
"grad_norm": 2.475109331033288,
"learning_rate": 1.8680477170809572e-06,
"loss": 0.9933,
"step": 169
},
{
"epoch": 0.4176904176904177,
"grad_norm": 1.4136668445270784,
"learning_rate": 1.8660254037844386e-06,
"loss": 0.9287,
"step": 170
},
{
"epoch": 0.4201474201474201,
"grad_norm": 1.7485785437438173,
"learning_rate": 1.863988822988634e-06,
"loss": 0.9749,
"step": 171
},
{
"epoch": 0.4226044226044226,
"grad_norm": 1.825009421972696,
"learning_rate": 1.8619380082455796e-06,
"loss": 0.9637,
"step": 172
},
{
"epoch": 0.4250614250614251,
"grad_norm": 1.6686645342967068,
"learning_rate": 1.85987299334181e-06,
"loss": 0.9951,
"step": 173
},
{
"epoch": 0.4275184275184275,
"grad_norm": 1.9731456114498873,
"learning_rate": 1.857793812297804e-06,
"loss": 0.9426,
"step": 174
},
{
"epoch": 0.42997542997543,
"grad_norm": 1.5558317172807916,
"learning_rate": 1.855700499367423e-06,
"loss": 0.9787,
"step": 175
},
{
"epoch": 0.43243243243243246,
"grad_norm": 2.098470725111635,
"learning_rate": 1.8535930890373463e-06,
"loss": 0.9936,
"step": 176
},
{
"epoch": 0.4348894348894349,
"grad_norm": 1.525785332653322,
"learning_rate": 1.8514716160265044e-06,
"loss": 0.9727,
"step": 177
},
{
"epoch": 0.43734643734643736,
"grad_norm": 1.4913555879508107,
"learning_rate": 1.8493361152855056e-06,
"loss": 0.9748,
"step": 178
},
{
"epoch": 0.4398034398034398,
"grad_norm": 1.340195087338962,
"learning_rate": 1.84718662199606e-06,
"loss": 0.9706,
"step": 179
},
{
"epoch": 0.44226044226044225,
"grad_norm": 2.5585128719755867,
"learning_rate": 1.8450231715704026e-06,
"loss": 0.9467,
"step": 180
},
{
"epoch": 0.44471744471744473,
"grad_norm": 1.3066387641923265,
"learning_rate": 1.8428457996507053e-06,
"loss": 0.9194,
"step": 181
},
{
"epoch": 0.44717444717444715,
"grad_norm": 1.6465015301936612,
"learning_rate": 1.8406545421084938e-06,
"loss": 1.0045,
"step": 182
},
{
"epoch": 0.44963144963144963,
"grad_norm": 1.4122218928192685,
"learning_rate": 1.8384494350440552e-06,
"loss": 0.9596,
"step": 183
},
{
"epoch": 0.4520884520884521,
"grad_norm": 1.4364915893717223,
"learning_rate": 1.8362305147858427e-06,
"loss": 0.9375,
"step": 184
},
{
"epoch": 0.45454545454545453,
"grad_norm": 1.5320964201665865,
"learning_rate": 1.833997817889878e-06,
"loss": 0.98,
"step": 185
},
{
"epoch": 0.457002457002457,
"grad_norm": 1.8963928416673594,
"learning_rate": 1.8317513811391477e-06,
"loss": 0.9499,
"step": 186
},
{
"epoch": 0.4594594594594595,
"grad_norm": 1.4450513723222074,
"learning_rate": 1.8294912415429993e-06,
"loss": 0.9387,
"step": 187
},
{
"epoch": 0.4619164619164619,
"grad_norm": 1.4227834836237356,
"learning_rate": 1.8272174363365298e-06,
"loss": 0.9856,
"step": 188
},
{
"epoch": 0.4643734643734644,
"grad_norm": 1.4225544869881186,
"learning_rate": 1.8249300029799732e-06,
"loss": 0.9201,
"step": 189
},
{
"epoch": 0.4668304668304668,
"grad_norm": 1.6580547740077536,
"learning_rate": 1.822628979158083e-06,
"loss": 0.9746,
"step": 190
},
{
"epoch": 0.4692874692874693,
"grad_norm": 1.6225782675239633,
"learning_rate": 1.8203144027795108e-06,
"loss": 0.9493,
"step": 191
},
{
"epoch": 0.47174447174447176,
"grad_norm": 1.8558149480612793,
"learning_rate": 1.8179863119761834e-06,
"loss": 0.9551,
"step": 192
},
{
"epoch": 0.4742014742014742,
"grad_norm": 1.5948966144212158,
"learning_rate": 1.815644745102673e-06,
"loss": 0.9799,
"step": 193
},
{
"epoch": 0.47665847665847666,
"grad_norm": 1.464419249562776,
"learning_rate": 1.8132897407355654e-06,
"loss": 0.9634,
"step": 194
},
{
"epoch": 0.47911547911547914,
"grad_norm": 1.6709063152230734,
"learning_rate": 1.8109213376728257e-06,
"loss": 0.977,
"step": 195
},
{
"epoch": 0.48157248157248156,
"grad_norm": 1.4980373204140025,
"learning_rate": 1.8085395749331579e-06,
"loss": 0.9839,
"step": 196
},
{
"epoch": 0.48402948402948404,
"grad_norm": 1.4728568491372855,
"learning_rate": 1.8061444917553627e-06,
"loss": 0.9791,
"step": 197
},
{
"epoch": 0.4864864864864865,
"grad_norm": 1.3529262817353658,
"learning_rate": 1.803736127597691e-06,
"loss": 0.9534,
"step": 198
},
{
"epoch": 0.48894348894348894,
"grad_norm": 1.4726330094781561,
"learning_rate": 1.8013145221371935e-06,
"loss": 0.9397,
"step": 199
},
{
"epoch": 0.4914004914004914,
"grad_norm": 1.4468557644109676,
"learning_rate": 1.7988797152690668e-06,
"loss": 0.8993,
"step": 200
},
{
"epoch": 0.49385749385749383,
"grad_norm": 1.39576948942371,
"learning_rate": 1.796431747105998e-06,
"loss": 0.945,
"step": 201
},
{
"epoch": 0.4963144963144963,
"grad_norm": 1.4248364851440443,
"learning_rate": 1.793970657977501e-06,
"loss": 0.9669,
"step": 202
},
{
"epoch": 0.4987714987714988,
"grad_norm": 1.4850571599264226,
"learning_rate": 1.791496488429254e-06,
"loss": 0.9572,
"step": 203
},
{
"epoch": 0.5012285012285013,
"grad_norm": 1.4619304758472125,
"learning_rate": 1.7890092792224313e-06,
"loss": 0.9604,
"step": 204
},
{
"epoch": 0.5036855036855037,
"grad_norm": 1.6208674632812228,
"learning_rate": 1.786509071333031e-06,
"loss": 0.9577,
"step": 205
},
{
"epoch": 0.5061425061425061,
"grad_norm": 1.4904987175582993,
"learning_rate": 1.7839959059512013e-06,
"loss": 0.9039,
"step": 206
},
{
"epoch": 0.5085995085995086,
"grad_norm": 1.4636256072722436,
"learning_rate": 1.7814698244805601e-06,
"loss": 0.9481,
"step": 207
},
{
"epoch": 0.5110565110565111,
"grad_norm": 1.3573990171789403,
"learning_rate": 1.7789308685375146e-06,
"loss": 0.9531,
"step": 208
},
{
"epoch": 0.5135135135135135,
"grad_norm": 1.4176202461890393,
"learning_rate": 1.7763790799505745e-06,
"loss": 0.959,
"step": 209
},
{
"epoch": 0.515970515970516,
"grad_norm": 1.5707837841315286,
"learning_rate": 1.773814500759663e-06,
"loss": 0.9798,
"step": 210
},
{
"epoch": 0.5184275184275184,
"grad_norm": 1.4706701895149787,
"learning_rate": 1.7712371732154256e-06,
"loss": 0.9928,
"step": 211
},
{
"epoch": 0.5208845208845209,
"grad_norm": 2.069991518090132,
"learning_rate": 1.7686471397785318e-06,
"loss": 1.0153,
"step": 212
},
{
"epoch": 0.5233415233415234,
"grad_norm": 1.443867260475285,
"learning_rate": 1.766044443118978e-06,
"loss": 0.9452,
"step": 213
},
{
"epoch": 0.5257985257985258,
"grad_norm": 1.5151736278472239,
"learning_rate": 1.7634291261153817e-06,
"loss": 0.9114,
"step": 214
},
{
"epoch": 0.5282555282555282,
"grad_norm": 1.3847804036493465,
"learning_rate": 1.7608012318542777e-06,
"loss": 0.9144,
"step": 215
},
{
"epoch": 0.5307125307125307,
"grad_norm": 1.415085937475802,
"learning_rate": 1.7581608036294075e-06,
"loss": 0.9726,
"step": 216
},
{
"epoch": 0.5331695331695332,
"grad_norm": 1.558702474871841,
"learning_rate": 1.7555078849410042e-06,
"loss": 0.9243,
"step": 217
},
{
"epoch": 0.5356265356265356,
"grad_norm": 1.826623068304597,
"learning_rate": 1.7528425194950793e-06,
"loss": 1.0106,
"step": 218
},
{
"epoch": 0.538083538083538,
"grad_norm": 1.5238015749509666,
"learning_rate": 1.7501647512026994e-06,
"loss": 1.0093,
"step": 219
},
{
"epoch": 0.5405405405405406,
"grad_norm": 1.6514187408836531,
"learning_rate": 1.7474746241792646e-06,
"loss": 0.991,
"step": 220
},
{
"epoch": 0.542997542997543,
"grad_norm": 1.5401870547381438,
"learning_rate": 1.7447721827437819e-06,
"loss": 0.8904,
"step": 221
},
{
"epoch": 0.5454545454545454,
"grad_norm": 1.58085107300811,
"learning_rate": 1.742057471418133e-06,
"loss": 0.907,
"step": 222
},
{
"epoch": 0.547911547911548,
"grad_norm": 1.4376088351554372,
"learning_rate": 1.7393305349263432e-06,
"loss": 0.9369,
"step": 223
},
{
"epoch": 0.5503685503685504,
"grad_norm": 1.3807320173451696,
"learning_rate": 1.7365914181938437e-06,
"loss": 0.8945,
"step": 224
},
{
"epoch": 0.5528255528255528,
"grad_norm": 1.5104969018841128,
"learning_rate": 1.7338401663467309e-06,
"loss": 0.9537,
"step": 225
},
{
"epoch": 0.5552825552825553,
"grad_norm": 1.4036248703567802,
"learning_rate": 1.7310768247110232e-06,
"loss": 0.9112,
"step": 226
},
{
"epoch": 0.5577395577395577,
"grad_norm": 1.4153113695817048,
"learning_rate": 1.7283014388119158e-06,
"loss": 0.9285,
"step": 227
},
{
"epoch": 0.5601965601965602,
"grad_norm": 1.6827361424168579,
"learning_rate": 1.7255140543730282e-06,
"loss": 0.9368,
"step": 228
},
{
"epoch": 0.5626535626535627,
"grad_norm": 1.52442672792253,
"learning_rate": 1.7227147173156523e-06,
"loss": 0.9838,
"step": 229
},
{
"epoch": 0.5651105651105651,
"grad_norm": 1.8631309960035536,
"learning_rate": 1.7199034737579958e-06,
"loss": 0.9311,
"step": 230
},
{
"epoch": 0.5675675675675675,
"grad_norm": 1.38816407656762,
"learning_rate": 1.7170803700144224e-06,
"loss": 0.945,
"step": 231
},
{
"epoch": 0.5700245700245701,
"grad_norm": 1.5370320107384547,
"learning_rate": 1.7142454525946889e-06,
"loss": 0.955,
"step": 232
},
{
"epoch": 0.5724815724815725,
"grad_norm": 1.5632251869407223,
"learning_rate": 1.7113987682031778e-06,
"loss": 0.9384,
"step": 233
},
{
"epoch": 0.5749385749385749,
"grad_norm": 1.602474983632772,
"learning_rate": 1.70854036373813e-06,
"loss": 0.992,
"step": 234
},
{
"epoch": 0.5773955773955773,
"grad_norm": 1.3375523516298262,
"learning_rate": 1.70567028629087e-06,
"loss": 0.941,
"step": 235
},
{
"epoch": 0.5798525798525799,
"grad_norm": 1.539702405955399,
"learning_rate": 1.7027885831450317e-06,
"loss": 0.9516,
"step": 236
},
{
"epoch": 0.5823095823095823,
"grad_norm": 1.6116292160800432,
"learning_rate": 1.6998953017757784e-06,
"loss": 0.992,
"step": 237
},
{
"epoch": 0.5847665847665847,
"grad_norm": 2.3507568939715835,
"learning_rate": 1.6969904898490213e-06,
"loss": 0.9165,
"step": 238
},
{
"epoch": 0.5872235872235873,
"grad_norm": 2.847402959001731,
"learning_rate": 1.694074195220634e-06,
"loss": 0.9985,
"step": 239
},
{
"epoch": 0.5896805896805897,
"grad_norm": 1.5211685594705786,
"learning_rate": 1.6911464659356632e-06,
"loss": 0.9365,
"step": 240
},
{
"epoch": 0.5921375921375921,
"grad_norm": 2.034636443003393,
"learning_rate": 1.688207350227539e-06,
"loss": 0.9094,
"step": 241
},
{
"epoch": 0.5945945945945946,
"grad_norm": 1.8768465174864484,
"learning_rate": 1.6852568965172791e-06,
"loss": 0.9898,
"step": 242
},
{
"epoch": 0.597051597051597,
"grad_norm": 1.7384412195063934,
"learning_rate": 1.6822951534126908e-06,
"loss": 0.9955,
"step": 243
},
{
"epoch": 0.5995085995085995,
"grad_norm": 1.4908111245385776,
"learning_rate": 1.6793221697075715e-06,
"loss": 0.9503,
"step": 244
},
{
"epoch": 0.601965601965602,
"grad_norm": 1.3104233864174921,
"learning_rate": 1.6763379943809026e-06,
"loss": 0.9087,
"step": 245
},
{
"epoch": 0.6044226044226044,
"grad_norm": 1.8378209471753764,
"learning_rate": 1.6733426765960456e-06,
"loss": 0.9295,
"step": 246
},
{
"epoch": 0.6068796068796068,
"grad_norm": 1.5513249549838153,
"learning_rate": 1.6703362656999299e-06,
"loss": 0.9622,
"step": 247
},
{
"epoch": 0.6093366093366094,
"grad_norm": 1.950619315682982,
"learning_rate": 1.6673188112222395e-06,
"loss": 0.9114,
"step": 248
},
{
"epoch": 0.6117936117936118,
"grad_norm": 1.5443899853661804,
"learning_rate": 1.6642903628745993e-06,
"loss": 0.9385,
"step": 249
},
{
"epoch": 0.6142506142506142,
"grad_norm": 1.842691343791668,
"learning_rate": 1.6612509705497541e-06,
"loss": 0.9681,
"step": 250
},
{
"epoch": 0.6167076167076168,
"grad_norm": 1.4579884817325401,
"learning_rate": 1.6582006843207479e-06,
"loss": 0.9967,
"step": 251
},
{
"epoch": 0.6191646191646192,
"grad_norm": 1.602595851064425,
"learning_rate": 1.655139554440098e-06,
"loss": 0.9674,
"step": 252
},
{
"epoch": 0.6216216216216216,
"grad_norm": 1.394530716544746,
"learning_rate": 1.652067631338967e-06,
"loss": 0.907,
"step": 253
},
{
"epoch": 0.6240786240786241,
"grad_norm": 1.6411505351669575,
"learning_rate": 1.6489849656263334e-06,
"loss": 0.9602,
"step": 254
},
{
"epoch": 0.6265356265356266,
"grad_norm": 1.7684281051463475,
"learning_rate": 1.6458916080881565e-06,
"loss": 1.0141,
"step": 255
},
{
"epoch": 0.628992628992629,
"grad_norm": 1.771327098736445,
"learning_rate": 1.6427876096865393e-06,
"loss": 0.9917,
"step": 256
},
{
"epoch": 0.6314496314496314,
"grad_norm": 1.440535084208374,
"learning_rate": 1.6396730215588912e-06,
"loss": 0.9326,
"step": 257
},
{
"epoch": 0.6339066339066339,
"grad_norm": 1.5197592107480333,
"learning_rate": 1.6365478950170833e-06,
"loss": 0.9475,
"step": 258
},
{
"epoch": 0.6363636363636364,
"grad_norm": 1.706431633980188,
"learning_rate": 1.6334122815466031e-06,
"loss": 0.9351,
"step": 259
},
{
"epoch": 0.6388206388206388,
"grad_norm": 1.323673519490258,
"learning_rate": 1.6302662328057085e-06,
"loss": 0.911,
"step": 260
},
{
"epoch": 0.6412776412776413,
"grad_norm": 1.5394197428392233,
"learning_rate": 1.627109800624574e-06,
"loss": 0.9812,
"step": 261
},
{
"epoch": 0.6437346437346437,
"grad_norm": 1.4717326057521631,
"learning_rate": 1.6239430370044386e-06,
"loss": 0.8713,
"step": 262
},
{
"epoch": 0.6461916461916462,
"grad_norm": 1.5141060178146766,
"learning_rate": 1.6207659941167484e-06,
"loss": 0.934,
"step": 263
},
{
"epoch": 0.6486486486486487,
"grad_norm": 1.606823271595554,
"learning_rate": 1.617578724302297e-06,
"loss": 0.9689,
"step": 264
},
{
"epoch": 0.6511056511056511,
"grad_norm": 1.436173862475149,
"learning_rate": 1.6143812800703643e-06,
"loss": 0.9524,
"step": 265
},
{
"epoch": 0.6535626535626535,
"grad_norm": 1.6794632700616086,
"learning_rate": 1.6111737140978491e-06,
"loss": 0.9377,
"step": 266
},
{
"epoch": 0.6560196560196561,
"grad_norm": 1.481170005860829,
"learning_rate": 1.6079560792284045e-06,
"loss": 0.9718,
"step": 267
},
{
"epoch": 0.6584766584766585,
"grad_norm": 1.5488863504284036,
"learning_rate": 1.604728428471564e-06,
"loss": 0.9597,
"step": 268
},
{
"epoch": 0.6609336609336609,
"grad_norm": 1.5751377061062102,
"learning_rate": 1.6014908150018703e-06,
"loss": 0.9741,
"step": 269
},
{
"epoch": 0.6633906633906634,
"grad_norm": 1.8133205336540266,
"learning_rate": 1.598243292157999e-06,
"loss": 0.9234,
"step": 270
},
{
"epoch": 0.6658476658476659,
"grad_norm": 2.014347908541205,
"learning_rate": 1.5949859134418797e-06,
"loss": 0.976,
"step": 271
},
{
"epoch": 0.6683046683046683,
"grad_norm": 1.3995748329733648,
"learning_rate": 1.5917187325178136e-06,
"loss": 0.9182,
"step": 272
},
{
"epoch": 0.6707616707616708,
"grad_norm": 1.536609373517338,
"learning_rate": 1.5884418032115906e-06,
"loss": 0.8921,
"step": 273
},
{
"epoch": 0.6732186732186732,
"grad_norm": 1.7529235851864882,
"learning_rate": 1.5851551795096026e-06,
"loss": 0.9083,
"step": 274
},
{
"epoch": 0.6756756756756757,
"grad_norm": 1.5263377609284405,
"learning_rate": 1.5818589155579529e-06,
"loss": 0.921,
"step": 275
},
{
"epoch": 0.6781326781326781,
"grad_norm": 1.841830762323041,
"learning_rate": 1.5785530656615652e-06,
"loss": 0.943,
"step": 276
},
{
"epoch": 0.6805896805896806,
"grad_norm": 1.5135421099319812,
"learning_rate": 1.5752376842832898e-06,
"loss": 0.9315,
"step": 277
},
{
"epoch": 0.683046683046683,
"grad_norm": 1.492488292559091,
"learning_rate": 1.571912826043003e-06,
"loss": 0.891,
"step": 278
},
{
"epoch": 0.6855036855036855,
"grad_norm": 1.9836354455315945,
"learning_rate": 1.5685785457167113e-06,
"loss": 0.8706,
"step": 279
},
{
"epoch": 0.687960687960688,
"grad_norm": 1.5134741374320924,
"learning_rate": 1.565234898235646e-06,
"loss": 0.9786,
"step": 280
},
{
"epoch": 0.6904176904176904,
"grad_norm": 1.4859838259301754,
"learning_rate": 1.5618819386853603e-06,
"loss": 0.9595,
"step": 281
},
{
"epoch": 0.6928746928746928,
"grad_norm": 1.567447965101693,
"learning_rate": 1.5585197223048202e-06,
"loss": 0.9373,
"step": 282
},
{
"epoch": 0.6953316953316954,
"grad_norm": 1.361598883424366,
"learning_rate": 1.5551483044854952e-06,
"loss": 0.9697,
"step": 283
},
{
"epoch": 0.6977886977886978,
"grad_norm": 1.6640657059750952,
"learning_rate": 1.551767740770446e-06,
"loss": 0.9244,
"step": 284
},
{
"epoch": 0.7002457002457002,
"grad_norm": 1.362536522369026,
"learning_rate": 1.5483780868534083e-06,
"loss": 0.981,
"step": 285
},
{
"epoch": 0.7027027027027027,
"grad_norm": 1.6010684699844948,
"learning_rate": 1.544979398577877e-06,
"loss": 1.0169,
"step": 286
},
{
"epoch": 0.7051597051597052,
"grad_norm": 1.6122664157541595,
"learning_rate": 1.5415717319361845e-06,
"loss": 0.9507,
"step": 287
},
{
"epoch": 0.7076167076167076,
"grad_norm": 1.5044459630841276,
"learning_rate": 1.5381551430685796e-06,
"loss": 0.9431,
"step": 288
},
{
"epoch": 0.7100737100737101,
"grad_norm": 1.464490997097453,
"learning_rate": 1.5347296882623016e-06,
"loss": 0.9726,
"step": 289
},
{
"epoch": 0.7125307125307125,
"grad_norm": 1.8674476124140849,
"learning_rate": 1.5312954239506533e-06,
"loss": 0.9406,
"step": 290
},
{
"epoch": 0.714987714987715,
"grad_norm": 1.3556749625191538,
"learning_rate": 1.5278524067120718e-06,
"loss": 0.8896,
"step": 291
},
{
"epoch": 0.7174447174447175,
"grad_norm": 1.4118443972410692,
"learning_rate": 1.5244006932691954e-06,
"loss": 0.9722,
"step": 292
},
{
"epoch": 0.7199017199017199,
"grad_norm": 1.6225004312255737,
"learning_rate": 1.5209403404879303e-06,
"loss": 0.929,
"step": 293
},
{
"epoch": 0.7223587223587223,
"grad_norm": 1.6022157008964288,
"learning_rate": 1.5174714053765123e-06,
"loss": 0.9497,
"step": 294
},
{
"epoch": 0.7248157248157249,
"grad_norm": 1.5891302329335815,
"learning_rate": 1.5139939450845699e-06,
"loss": 0.973,
"step": 295
},
{
"epoch": 0.7272727272727273,
"grad_norm": 1.6988844584422054,
"learning_rate": 1.5105080169021789e-06,
"loss": 0.9184,
"step": 296
},
{
"epoch": 0.7297297297297297,
"grad_norm": 1.453985001696748,
"learning_rate": 1.5070136782589234e-06,
"loss": 0.9477,
"step": 297
},
{
"epoch": 0.7321867321867321,
"grad_norm": 1.604839180071105,
"learning_rate": 1.5035109867229457e-06,
"loss": 0.9409,
"step": 298
},
{
"epoch": 0.7346437346437347,
"grad_norm": 1.5773771725228862,
"learning_rate": 1.5e-06,
"loss": 0.937,
"step": 299
},
{
"epoch": 0.7371007371007371,
"grad_norm": 1.4272988880705397,
"learning_rate": 1.496480775932501e-06,
"loss": 0.9481,
"step": 300
},
{
"epoch": 0.7395577395577395,
"grad_norm": 1.534183187925811,
"learning_rate": 1.492953372498571e-06,
"loss": 0.9111,
"step": 301
},
{
"epoch": 0.742014742014742,
"grad_norm": 1.6285985984253928,
"learning_rate": 1.4894178478110854e-06,
"loss": 0.949,
"step": 302
},
{
"epoch": 0.7444717444717445,
"grad_norm": 1.521782496026601,
"learning_rate": 1.4858742601167138e-06,
"loss": 0.9405,
"step": 303
},
{
"epoch": 0.7469287469287469,
"grad_norm": 1.522930763087887,
"learning_rate": 1.4823226677949622e-06,
"loss": 0.949,
"step": 304
},
{
"epoch": 0.7493857493857494,
"grad_norm": 5.354032445168916,
"learning_rate": 1.4787631293572092e-06,
"loss": 0.9708,
"step": 305
},
{
"epoch": 0.7518427518427518,
"grad_norm": 1.199402404810751,
"learning_rate": 1.4751957034457446e-06,
"loss": 0.8895,
"step": 306
},
{
"epoch": 0.7542997542997543,
"grad_norm": 1.4889622591506924,
"learning_rate": 1.4716204488328005e-06,
"loss": 0.9258,
"step": 307
},
{
"epoch": 0.7567567567567568,
"grad_norm": 1.586929861427342,
"learning_rate": 1.4680374244195858e-06,
"loss": 0.9523,
"step": 308
},
{
"epoch": 0.7592137592137592,
"grad_norm": 1.7542006852572127,
"learning_rate": 1.464446689235314e-06,
"loss": 0.9671,
"step": 309
},
{
"epoch": 0.7616707616707616,
"grad_norm": 1.906886186477315,
"learning_rate": 1.4608483024362302e-06,
"loss": 0.9479,
"step": 310
},
{
"epoch": 0.7641277641277642,
"grad_norm": 1.3426576295326547,
"learning_rate": 1.4572423233046385e-06,
"loss": 0.9245,
"step": 311
},
{
"epoch": 0.7665847665847666,
"grad_norm": 1.544123812480689,
"learning_rate": 1.453628811247924e-06,
"loss": 0.9833,
"step": 312
},
{
"epoch": 0.769041769041769,
"grad_norm": 1.3070207560130556,
"learning_rate": 1.4500078257975745e-06,
"loss": 0.9518,
"step": 313
},
{
"epoch": 0.7714987714987716,
"grad_norm": 1.953852840958092,
"learning_rate": 1.4463794266081992e-06,
"loss": 0.8871,
"step": 314
},
{
"epoch": 0.773955773955774,
"grad_norm": 1.5767416529599687,
"learning_rate": 1.4427436734565473e-06,
"loss": 0.9921,
"step": 315
},
{
"epoch": 0.7764127764127764,
"grad_norm": 1.5600472966901433,
"learning_rate": 1.4391006262405212e-06,
"loss": 0.9068,
"step": 316
},
{
"epoch": 0.7788697788697788,
"grad_norm": 1.8524989491255999,
"learning_rate": 1.4354503449781913e-06,
"loss": 0.8956,
"step": 317
},
{
"epoch": 0.7813267813267813,
"grad_norm": 1.4701804028885344,
"learning_rate": 1.4317928898068066e-06,
"loss": 0.9547,
"step": 318
},
{
"epoch": 0.7837837837837838,
"grad_norm": 1.505304585103799,
"learning_rate": 1.4281283209818039e-06,
"loss": 0.9214,
"step": 319
},
{
"epoch": 0.7862407862407862,
"grad_norm": 1.5458422714438642,
"learning_rate": 1.424456698875815e-06,
"loss": 0.9711,
"step": 320
},
{
"epoch": 0.7886977886977887,
"grad_norm": 1.8126365786822831,
"learning_rate": 1.4207780839776732e-06,
"loss": 0.9638,
"step": 321
},
{
"epoch": 0.7911547911547911,
"grad_norm": 1.3634824425424106,
"learning_rate": 1.4170925368914151e-06,
"loss": 0.925,
"step": 322
},
{
"epoch": 0.7936117936117936,
"grad_norm": 1.5523412959213532,
"learning_rate": 1.413400118335283e-06,
"loss": 0.9681,
"step": 323
},
{
"epoch": 0.7960687960687961,
"grad_norm": 1.4281815228481956,
"learning_rate": 1.4097008891407245e-06,
"loss": 0.9469,
"step": 324
},
{
"epoch": 0.7985257985257985,
"grad_norm": 1.755464439866289,
"learning_rate": 1.4059949102513913e-06,
"loss": 0.9614,
"step": 325
},
{
"epoch": 0.800982800982801,
"grad_norm": 1.3755915444522364,
"learning_rate": 1.4022822427221323e-06,
"loss": 0.9344,
"step": 326
},
{
"epoch": 0.8034398034398035,
"grad_norm": 1.6217750884199984,
"learning_rate": 1.3985629477179915e-06,
"loss": 0.9499,
"step": 327
},
{
"epoch": 0.8058968058968059,
"grad_norm": 1.5965334157340827,
"learning_rate": 1.3948370865131978e-06,
"loss": 0.9433,
"step": 328
},
{
"epoch": 0.8083538083538083,
"grad_norm": 1.8371436403022903,
"learning_rate": 1.3911047204901558e-06,
"loss": 0.9899,
"step": 329
},
{
"epoch": 0.8108108108108109,
"grad_norm": 1.8839520322523513,
"learning_rate": 1.387365911138436e-06,
"loss": 0.9664,
"step": 330
},
{
"epoch": 0.8132678132678133,
"grad_norm": 1.495735998773526,
"learning_rate": 1.3836207200537594e-06,
"loss": 0.9784,
"step": 331
},
{
"epoch": 0.8157248157248157,
"grad_norm": 1.4893770976659941,
"learning_rate": 1.3798692089369855e-06,
"loss": 0.9046,
"step": 332
},
{
"epoch": 0.8181818181818182,
"grad_norm": 1.5373781226372853,
"learning_rate": 1.3761114395930928e-06,
"loss": 0.9461,
"step": 333
},
{
"epoch": 0.8206388206388207,
"grad_norm": 1.497679579290753,
"learning_rate": 1.3723474739301634e-06,
"loss": 0.9109,
"step": 334
},
{
"epoch": 0.8230958230958231,
"grad_norm": 1.6333646657141205,
"learning_rate": 1.3685773739583618e-06,
"loss": 0.893,
"step": 335
},
{
"epoch": 0.8255528255528255,
"grad_norm": 1.4444606139573457,
"learning_rate": 1.3648012017889121e-06,
"loss": 0.9616,
"step": 336
},
{
"epoch": 0.828009828009828,
"grad_norm": 1.633866194720376,
"learning_rate": 1.3610190196330775e-06,
"loss": 0.9316,
"step": 337
},
{
"epoch": 0.8304668304668305,
"grad_norm": 1.4965633625059045,
"learning_rate": 1.3572308898011327e-06,
"loss": 0.8995,
"step": 338
},
{
"epoch": 0.8329238329238329,
"grad_norm": 1.4593159509482825,
"learning_rate": 1.3534368747013394e-06,
"loss": 0.9416,
"step": 339
},
{
"epoch": 0.8353808353808354,
"grad_norm": 1.5060565758495885,
"learning_rate": 1.3496370368389165e-06,
"loss": 0.9385,
"step": 340
},
{
"epoch": 0.8378378378378378,
"grad_norm": 1.5228727270675948,
"learning_rate": 1.3458314388150113e-06,
"loss": 0.9249,
"step": 341
},
{
"epoch": 0.8402948402948403,
"grad_norm": 1.6137336621038334,
"learning_rate": 1.3420201433256689e-06,
"loss": 0.9549,
"step": 342
},
{
"epoch": 0.8427518427518428,
"grad_norm": 1.2744326108891244,
"learning_rate": 1.3382032131607966e-06,
"loss": 0.9275,
"step": 343
},
{
"epoch": 0.8452088452088452,
"grad_norm": 1.5480593545772812,
"learning_rate": 1.3343807112031326e-06,
"loss": 0.9342,
"step": 344
},
{
"epoch": 0.8476658476658476,
"grad_norm": 1.6690119423341088,
"learning_rate": 1.3305527004272087e-06,
"loss": 0.9375,
"step": 345
},
{
"epoch": 0.8501228501228502,
"grad_norm": 1.4907525884430786,
"learning_rate": 1.3267192438983117e-06,
"loss": 0.9183,
"step": 346
},
{
"epoch": 0.8525798525798526,
"grad_norm": 1.487251447150466,
"learning_rate": 1.3228804047714462e-06,
"loss": 0.9572,
"step": 347
},
{
"epoch": 0.855036855036855,
"grad_norm": 2.022254168777352,
"learning_rate": 1.3190362462902936e-06,
"loss": 0.9431,
"step": 348
},
{
"epoch": 0.8574938574938575,
"grad_norm": 1.570248437527878,
"learning_rate": 1.31518683178617e-06,
"loss": 0.8945,
"step": 349
},
{
"epoch": 0.85995085995086,
"grad_norm": 1.5017196199325449,
"learning_rate": 1.3113322246769816e-06,
"loss": 0.8841,
"step": 350
},
{
"epoch": 0.8624078624078624,
"grad_norm": 1.5003226831400922,
"learning_rate": 1.3074724884661831e-06,
"loss": 0.9785,
"step": 351
},
{
"epoch": 0.8648648648648649,
"grad_norm": 1.6237720015743136,
"learning_rate": 1.3036076867417286e-06,
"loss": 0.9942,
"step": 352
},
{
"epoch": 0.8673218673218673,
"grad_norm": 1.586236748920305,
"learning_rate": 1.299737883175024e-06,
"loss": 0.9036,
"step": 353
},
{
"epoch": 0.8697788697788698,
"grad_norm": 1.6356021082940453,
"learning_rate": 1.2958631415198813e-06,
"loss": 0.9723,
"step": 354
},
{
"epoch": 0.8722358722358723,
"grad_norm": 1.5448677133530493,
"learning_rate": 1.2919835256114637e-06,
"loss": 0.9391,
"step": 355
},
{
"epoch": 0.8746928746928747,
"grad_norm": 1.6212696290821345,
"learning_rate": 1.2880990993652377e-06,
"loss": 0.9202,
"step": 356
},
{
"epoch": 0.8771498771498771,
"grad_norm": 1.5176857112402762,
"learning_rate": 1.2842099267759175e-06,
"loss": 0.9037,
"step": 357
},
{
"epoch": 0.8796068796068796,
"grad_norm": 1.3944467856106335,
"learning_rate": 1.2803160719164126e-06,
"loss": 0.8854,
"step": 358
},
{
"epoch": 0.8820638820638821,
"grad_norm": 1.4377067419239682,
"learning_rate": 1.2764175989367715e-06,
"loss": 0.905,
"step": 359
},
{
"epoch": 0.8845208845208845,
"grad_norm": 1.6123612736739008,
"learning_rate": 1.272514572063124e-06,
"loss": 0.9881,
"step": 360
},
{
"epoch": 0.8869778869778869,
"grad_norm": 1.589984246633212,
"learning_rate": 1.2686070555966253e-06,
"loss": 0.9529,
"step": 361
},
{
"epoch": 0.8894348894348895,
"grad_norm": 1.4049932596277341,
"learning_rate": 1.2646951139123933e-06,
"loss": 0.8715,
"step": 362
},
{
"epoch": 0.8918918918918919,
"grad_norm": 1.7376495814908663,
"learning_rate": 1.260778811458452e-06,
"loss": 0.9551,
"step": 363
},
{
"epoch": 0.8943488943488943,
"grad_norm": 1.6356081133105504,
"learning_rate": 1.2568582127546661e-06,
"loss": 0.9357,
"step": 364
},
{
"epoch": 0.8968058968058968,
"grad_norm": 1.5620481027616258,
"learning_rate": 1.2529333823916806e-06,
"loss": 0.9724,
"step": 365
},
{
"epoch": 0.8992628992628993,
"grad_norm": 1.678396238284589,
"learning_rate": 1.2490043850298557e-06,
"loss": 0.9156,
"step": 366
},
{
"epoch": 0.9017199017199017,
"grad_norm": 2.354758429687385,
"learning_rate": 1.2450712853982013e-06,
"loss": 0.8628,
"step": 367
},
{
"epoch": 0.9041769041769042,
"grad_norm": 1.600899191967723,
"learning_rate": 1.2411341482933108e-06,
"loss": 0.9431,
"step": 368
},
{
"epoch": 0.9066339066339066,
"grad_norm": 1.4058440155556784,
"learning_rate": 1.2371930385782943e-06,
"loss": 0.9446,
"step": 369
},
{
"epoch": 0.9090909090909091,
"grad_norm": 1.4613065350370973,
"learning_rate": 1.233248021181709e-06,
"loss": 0.9164,
"step": 370
},
{
"epoch": 0.9115479115479116,
"grad_norm": 1.6113469899034865,
"learning_rate": 1.22929916109649e-06,
"loss": 0.9209,
"step": 371
},
{
"epoch": 0.914004914004914,
"grad_norm": 1.4540218388463704,
"learning_rate": 1.2253465233788794e-06,
"loss": 0.9865,
"step": 372
},
{
"epoch": 0.9164619164619164,
"grad_norm": 1.4002046410996662,
"learning_rate": 1.2213901731473553e-06,
"loss": 0.9229,
"step": 373
},
{
"epoch": 0.918918918918919,
"grad_norm": 1.4545211228962136,
"learning_rate": 1.217430175581557e-06,
"loss": 0.9401,
"step": 374
},
{
"epoch": 0.9213759213759214,
"grad_norm": 1.4460543335595146,
"learning_rate": 1.2134665959212135e-06,
"loss": 0.9552,
"step": 375
},
{
"epoch": 0.9238329238329238,
"grad_norm": 1.5791517432252247,
"learning_rate": 1.209499499465068e-06,
"loss": 0.9547,
"step": 376
},
{
"epoch": 0.9262899262899262,
"grad_norm": 1.4523281683604505,
"learning_rate": 1.2055289515698006e-06,
"loss": 0.9669,
"step": 377
},
{
"epoch": 0.9287469287469288,
"grad_norm": 1.8584285735562196,
"learning_rate": 1.2015550176489537e-06,
"loss": 0.9434,
"step": 378
},
{
"epoch": 0.9312039312039312,
"grad_norm": 1.5086682042761248,
"learning_rate": 1.1975777631718531e-06,
"loss": 0.9377,
"step": 379
},
{
"epoch": 0.9336609336609336,
"grad_norm": 1.7455613963553367,
"learning_rate": 1.19359725366253e-06,
"loss": 0.919,
"step": 380
},
{
"epoch": 0.9361179361179361,
"grad_norm": 1.5307030954736855,
"learning_rate": 1.1896135546986409e-06,
"loss": 0.9282,
"step": 381
},
{
"epoch": 0.9385749385749386,
"grad_norm": 1.5159714020228543,
"learning_rate": 1.1856267319103876e-06,
"loss": 0.9765,
"step": 382
},
{
"epoch": 0.941031941031941,
"grad_norm": 1.5839779729377836,
"learning_rate": 1.1816368509794364e-06,
"loss": 0.9793,
"step": 383
},
{
"epoch": 0.9434889434889435,
"grad_norm": 1.3472301054596898,
"learning_rate": 1.177643977637835e-06,
"loss": 0.9189,
"step": 384
},
{
"epoch": 0.9459459459459459,
"grad_norm": 1.5340710636652473,
"learning_rate": 1.1736481776669305e-06,
"loss": 0.9663,
"step": 385
},
{
"epoch": 0.9484029484029484,
"grad_norm": 1.4047565697947708,
"learning_rate": 1.1696495168962846e-06,
"loss": 0.9207,
"step": 386
},
{
"epoch": 0.9508599508599509,
"grad_norm": 1.6035760956637677,
"learning_rate": 1.165648061202591e-06,
"loss": 0.9793,
"step": 387
},
{
"epoch": 0.9533169533169533,
"grad_norm": 1.4948457768549497,
"learning_rate": 1.1616438765085882e-06,
"loss": 0.9381,
"step": 388
},
{
"epoch": 0.9557739557739557,
"grad_norm": 1.5241351339551144,
"learning_rate": 1.1576370287819734e-06,
"loss": 0.931,
"step": 389
},
{
"epoch": 0.9582309582309583,
"grad_norm": 1.4995817174577337,
"learning_rate": 1.1536275840343183e-06,
"loss": 0.9594,
"step": 390
},
{
"epoch": 0.9606879606879607,
"grad_norm": 1.8781230050093483,
"learning_rate": 1.149615608319978e-06,
"loss": 0.9536,
"step": 391
},
{
"epoch": 0.9631449631449631,
"grad_norm": 1.5385341714632899,
"learning_rate": 1.145601167735005e-06,
"loss": 0.9281,
"step": 392
},
{
"epoch": 0.9656019656019657,
"grad_norm": 1.6183085918500024,
"learning_rate": 1.1415843284160597e-06,
"loss": 0.9366,
"step": 393
},
{
"epoch": 0.9680589680589681,
"grad_norm": 1.5534604794185942,
"learning_rate": 1.137565156539322e-06,
"loss": 0.9316,
"step": 394
},
{
"epoch": 0.9705159705159705,
"grad_norm": 1.423811752511575,
"learning_rate": 1.133543718319398e-06,
"loss": 0.8927,
"step": 395
},
{
"epoch": 0.972972972972973,
"grad_norm": 1.4578205842089746,
"learning_rate": 1.1295200800082325e-06,
"loss": 0.951,
"step": 396
},
{
"epoch": 0.9754299754299754,
"grad_norm": 1.6386231495575438,
"learning_rate": 1.125494307894016e-06,
"loss": 0.8952,
"step": 397
},
{
"epoch": 0.9778869778869779,
"grad_norm": 1.5075389802303287,
"learning_rate": 1.1214664683000924e-06,
"loss": 0.8966,
"step": 398
},
{
"epoch": 0.9803439803439803,
"grad_norm": 1.631579827773725,
"learning_rate": 1.1174366275838662e-06,
"loss": 0.9411,
"step": 399
},
{
"epoch": 0.9828009828009828,
"grad_norm": 1.8249088799770508,
"learning_rate": 1.1134048521357116e-06,
"loss": 0.9232,
"step": 400
},
{
"epoch": 0.9852579852579852,
"grad_norm": 1.4813873308103862,
"learning_rate": 1.1093712083778746e-06,
"loss": 0.9625,
"step": 401
},
{
"epoch": 0.9877149877149877,
"grad_norm": 1.5996089579510389,
"learning_rate": 1.105335762763382e-06,
"loss": 0.9592,
"step": 402
},
{
"epoch": 0.9901719901719902,
"grad_norm": 1.4507338601430246,
"learning_rate": 1.1012985817749462e-06,
"loss": 0.9558,
"step": 403
},
{
"epoch": 0.9926289926289926,
"grad_norm": 1.452275573847242,
"learning_rate": 1.097259731923869e-06,
"loss": 0.9582,
"step": 404
},
{
"epoch": 0.995085995085995,
"grad_norm": 1.4529835839819856,
"learning_rate": 1.0932192797489458e-06,
"loss": 0.8952,
"step": 405
},
{
"epoch": 0.9975429975429976,
"grad_norm": 2.622272111272181,
"learning_rate": 1.0891772918153694e-06,
"loss": 0.918,
"step": 406
},
{
"epoch": 1.0,
"grad_norm": 1.5285112861748582,
"learning_rate": 1.0851338347136355e-06,
"loss": 0.877,
"step": 407
},
{
"epoch": 1.0024570024570025,
"grad_norm": 1.4118845101420152,
"learning_rate": 1.0810889750584424e-06,
"loss": 0.8647,
"step": 408
},
{
"epoch": 1.0049140049140048,
"grad_norm": 1.6747623259684568,
"learning_rate": 1.0770427794875948e-06,
"loss": 0.9671,
"step": 409
},
{
"epoch": 1.0073710073710074,
"grad_norm": 1.307864675513283,
"learning_rate": 1.0729953146609075e-06,
"loss": 0.8296,
"step": 410
},
{
"epoch": 1.00982800982801,
"grad_norm": 1.4553500313995769,
"learning_rate": 1.0689466472591048e-06,
"loss": 0.937,
"step": 411
},
{
"epoch": 1.0122850122850122,
"grad_norm": 1.4315419855901317,
"learning_rate": 1.0648968439827239e-06,
"loss": 0.9042,
"step": 412
},
{
"epoch": 1.0147420147420148,
"grad_norm": 1.5739844573571937,
"learning_rate": 1.0608459715510139e-06,
"loss": 0.9011,
"step": 413
},
{
"epoch": 1.0171990171990173,
"grad_norm": 1.437687738713176,
"learning_rate": 1.0567940967008395e-06,
"loss": 0.9363,
"step": 414
},
{
"epoch": 1.0196560196560196,
"grad_norm": 1.534448861262034,
"learning_rate": 1.052741286185579e-06,
"loss": 0.9539,
"step": 415
},
{
"epoch": 1.0221130221130221,
"grad_norm": 1.644806928510893,
"learning_rate": 1.0486876067740252e-06,
"loss": 0.9802,
"step": 416
},
{
"epoch": 1.0245700245700247,
"grad_norm": 1.5874571507432869,
"learning_rate": 1.0446331252492864e-06,
"loss": 0.8751,
"step": 417
},
{
"epoch": 1.027027027027027,
"grad_norm": 1.5244813435420985,
"learning_rate": 1.0405779084076856e-06,
"loss": 0.8844,
"step": 418
},
{
"epoch": 1.0294840294840295,
"grad_norm": 1.5721376461339462,
"learning_rate": 1.036522023057659e-06,
"loss": 0.9593,
"step": 419
},
{
"epoch": 1.031941031941032,
"grad_norm": 3.006639344932307,
"learning_rate": 1.0324655360186567e-06,
"loss": 0.9055,
"step": 420
},
{
"epoch": 1.0343980343980343,
"grad_norm": 1.3782225003410025,
"learning_rate": 1.0284085141200423e-06,
"loss": 0.9183,
"step": 421
},
{
"epoch": 1.0368550368550369,
"grad_norm": 1.376873801999615,
"learning_rate": 1.0243510241999897e-06,
"loss": 0.8767,
"step": 422
},
{
"epoch": 1.0393120393120394,
"grad_norm": 1.394373200963149,
"learning_rate": 1.0202931331043839e-06,
"loss": 0.8821,
"step": 423
},
{
"epoch": 1.0417690417690417,
"grad_norm": 1.571320978169866,
"learning_rate": 1.0162349076857191e-06,
"loss": 0.8959,
"step": 424
},
{
"epoch": 1.0442260442260443,
"grad_norm": 1.5336886338120022,
"learning_rate": 1.0121764148019975e-06,
"loss": 0.9375,
"step": 425
},
{
"epoch": 1.0466830466830466,
"grad_norm": 1.4693963607748404,
"learning_rate": 1.0081177213156277e-06,
"loss": 0.8909,
"step": 426
},
{
"epoch": 1.049140049140049,
"grad_norm": 1.5830327984938106,
"learning_rate": 1.004058894092323e-06,
"loss": 0.9207,
"step": 427
},
{
"epoch": 1.0515970515970516,
"grad_norm": 1.5170501490428927,
"learning_rate": 1e-06,
"loss": 0.9036,
"step": 428
},
{
"epoch": 1.054054054054054,
"grad_norm": 1.5692066826348359,
"learning_rate": 9.95941105907677e-07,
"loss": 0.9193,
"step": 429
},
{
"epoch": 1.0565110565110565,
"grad_norm": 1.4357790543992526,
"learning_rate": 9.918822786843724e-07,
"loss": 0.9124,
"step": 430
},
{
"epoch": 1.058968058968059,
"grad_norm": 1.423440533928829,
"learning_rate": 9.878235851980024e-07,
"loss": 0.9146,
"step": 431
},
{
"epoch": 1.0614250614250613,
"grad_norm": 1.590468077725602,
"learning_rate": 9.837650923142808e-07,
"loss": 0.8705,
"step": 432
},
{
"epoch": 1.0638820638820639,
"grad_norm": 1.461231262115798,
"learning_rate": 9.797068668956163e-07,
"loss": 0.9438,
"step": 433
},
{
"epoch": 1.0663390663390664,
"grad_norm": 1.42661968764142,
"learning_rate": 9.756489758000105e-07,
"loss": 0.8747,
"step": 434
},
{
"epoch": 1.0687960687960687,
"grad_norm": 1.4657641576186988,
"learning_rate": 9.715914858799576e-07,
"loss": 0.8939,
"step": 435
},
{
"epoch": 1.0712530712530712,
"grad_norm": 1.5022324634860065,
"learning_rate": 9.675344639813434e-07,
"loss": 0.922,
"step": 436
},
{
"epoch": 1.0737100737100738,
"grad_norm": 1.424039088022573,
"learning_rate": 9.63477976942341e-07,
"loss": 0.953,
"step": 437
},
{
"epoch": 1.076167076167076,
"grad_norm": 1.5965077678963564,
"learning_rate": 9.594220915923147e-07,
"loss": 0.9443,
"step": 438
},
{
"epoch": 1.0786240786240786,
"grad_norm": 1.6371319904729476,
"learning_rate": 9.553668747507138e-07,
"loss": 0.9228,
"step": 439
},
{
"epoch": 1.0810810810810811,
"grad_norm": 1.8309314739878833,
"learning_rate": 9.51312393225975e-07,
"loss": 0.8753,
"step": 440
},
{
"epoch": 1.0835380835380835,
"grad_norm": 1.39319457517673,
"learning_rate": 9.472587138144215e-07,
"loss": 0.8939,
"step": 441
},
{
"epoch": 1.085995085995086,
"grad_norm": 1.5070197645260877,
"learning_rate": 9.432059032991607e-07,
"loss": 0.9123,
"step": 442
},
{
"epoch": 1.0884520884520885,
"grad_norm": 1.3753531136081873,
"learning_rate": 9.391540284489861e-07,
"loss": 0.9328,
"step": 443
},
{
"epoch": 1.0909090909090908,
"grad_norm": 1.4036108581529319,
"learning_rate": 9.351031560172764e-07,
"loss": 0.9138,
"step": 444
},
{
"epoch": 1.0933660933660934,
"grad_norm": 1.460079134056615,
"learning_rate": 9.310533527408951e-07,
"loss": 0.8736,
"step": 445
},
{
"epoch": 1.095823095823096,
"grad_norm": 2.1567900139463663,
"learning_rate": 9.270046853390924e-07,
"loss": 0.9649,
"step": 446
},
{
"epoch": 1.0982800982800982,
"grad_norm": 1.5086116320926348,
"learning_rate": 9.229572205124051e-07,
"loss": 0.8839,
"step": 447
},
{
"epoch": 1.1007371007371007,
"grad_norm": 1.4696338643312472,
"learning_rate": 9.189110249415576e-07,
"loss": 0.8971,
"step": 448
},
{
"epoch": 1.1031941031941033,
"grad_norm": 1.5799491675072612,
"learning_rate": 9.148661652863642e-07,
"loss": 0.8839,
"step": 449
},
{
"epoch": 1.1056511056511056,
"grad_norm": 1.633966629237029,
"learning_rate": 9.108227081846304e-07,
"loss": 0.9053,
"step": 450
},
{
"epoch": 1.1081081081081081,
"grad_norm": 1.5227318714591582,
"learning_rate": 9.067807202510542e-07,
"loss": 0.8734,
"step": 451
},
{
"epoch": 1.1105651105651106,
"grad_norm": 1.6390683121943366,
"learning_rate": 9.027402680761308e-07,
"loss": 0.8389,
"step": 452
},
{
"epoch": 1.113022113022113,
"grad_norm": 1.8480120093100039,
"learning_rate": 8.987014182250537e-07,
"loss": 0.8977,
"step": 453
},
{
"epoch": 1.1154791154791155,
"grad_norm": 1.6432113495263132,
"learning_rate": 8.94664237236618e-07,
"loss": 0.8924,
"step": 454
},
{
"epoch": 1.117936117936118,
"grad_norm": 1.7383339935167776,
"learning_rate": 8.906287916221258e-07,
"loss": 0.9181,
"step": 455
},
{
"epoch": 1.1203931203931203,
"grad_norm": 1.5888604077241295,
"learning_rate": 8.865951478642886e-07,
"loss": 0.9276,
"step": 456
},
{
"epoch": 1.1228501228501229,
"grad_norm": 1.4607976669981695,
"learning_rate": 8.825633724161335e-07,
"loss": 0.8802,
"step": 457
},
{
"epoch": 1.1253071253071254,
"grad_norm": 1.1981728862199128,
"learning_rate": 8.785335316999078e-07,
"loss": 0.9242,
"step": 458
},
{
"epoch": 1.1277641277641277,
"grad_norm": 1.5537897318688183,
"learning_rate": 8.745056921059839e-07,
"loss": 0.9001,
"step": 459
},
{
"epoch": 1.1302211302211302,
"grad_norm": 1.475963647033354,
"learning_rate": 8.704799199917672e-07,
"loss": 0.8775,
"step": 460
},
{
"epoch": 1.1326781326781328,
"grad_norm": 1.3982302407588918,
"learning_rate": 8.664562816806021e-07,
"loss": 0.8828,
"step": 461
},
{
"epoch": 1.135135135135135,
"grad_norm": 1.4892642936338247,
"learning_rate": 8.624348434606781e-07,
"loss": 0.9042,
"step": 462
},
{
"epoch": 1.1375921375921376,
"grad_norm": 1.8141339480004706,
"learning_rate": 8.584156715839401e-07,
"loss": 0.8748,
"step": 463
},
{
"epoch": 1.1400491400491402,
"grad_norm": 1.641438804529928,
"learning_rate": 8.543988322649953e-07,
"loss": 0.9006,
"step": 464
},
{
"epoch": 1.1425061425061425,
"grad_norm": 1.5215620001272283,
"learning_rate": 8.503843916800219e-07,
"loss": 0.9418,
"step": 465
},
{
"epoch": 1.144963144963145,
"grad_norm": 1.5095379213125475,
"learning_rate": 8.463724159656815e-07,
"loss": 0.9072,
"step": 466
},
{
"epoch": 1.1474201474201475,
"grad_norm": 1.4101920682297036,
"learning_rate": 8.423629712180265e-07,
"loss": 0.8929,
"step": 467
},
{
"epoch": 1.1498771498771498,
"grad_norm": 1.5017183949322925,
"learning_rate": 8.38356123491412e-07,
"loss": 0.9246,
"step": 468
},
{
"epoch": 1.1523341523341524,
"grad_norm": 1.4487713518305103,
"learning_rate": 8.34351938797409e-07,
"loss": 0.9978,
"step": 469
},
{
"epoch": 1.154791154791155,
"grad_norm": 1.9019229859778328,
"learning_rate": 8.303504831037153e-07,
"loss": 0.8963,
"step": 470
},
{
"epoch": 1.1572481572481572,
"grad_norm": 1.5553306743565323,
"learning_rate": 8.263518223330696e-07,
"loss": 0.9076,
"step": 471
},
{
"epoch": 1.1597051597051597,
"grad_norm": 1.528301521498108,
"learning_rate": 8.22356022362165e-07,
"loss": 0.9005,
"step": 472
},
{
"epoch": 1.1621621621621623,
"grad_norm": 1.4246755498283765,
"learning_rate": 8.183631490205635e-07,
"loss": 0.8751,
"step": 473
},
{
"epoch": 1.1646191646191646,
"grad_norm": 1.4837658920113785,
"learning_rate": 8.143732680896123e-07,
"loss": 0.8911,
"step": 474
},
{
"epoch": 1.1670761670761671,
"grad_norm": 1.3838573607006248,
"learning_rate": 8.103864453013592e-07,
"loss": 0.8752,
"step": 475
},
{
"epoch": 1.1695331695331694,
"grad_norm": 1.3956886722399857,
"learning_rate": 8.0640274633747e-07,
"loss": 0.907,
"step": 476
},
{
"epoch": 1.171990171990172,
"grad_norm": 1.4606205140333972,
"learning_rate": 8.024222368281468e-07,
"loss": 0.964,
"step": 477
},
{
"epoch": 1.1744471744471745,
"grad_norm": 1.438679071763127,
"learning_rate": 7.984449823510466e-07,
"loss": 0.8499,
"step": 478
},
{
"epoch": 1.1769041769041768,
"grad_norm": 1.6564632287349166,
"learning_rate": 7.944710484301994e-07,
"loss": 0.892,
"step": 479
},
{
"epoch": 1.1793611793611793,
"grad_norm": 1.5772645911948133,
"learning_rate": 7.905005005349319e-07,
"loss": 0.8451,
"step": 480
},
{
"epoch": 1.1818181818181819,
"grad_norm": 1.5408140168729385,
"learning_rate": 7.865334040787865e-07,
"loss": 0.9295,
"step": 481
},
{
"epoch": 1.1842751842751842,
"grad_norm": 1.5386445563903186,
"learning_rate": 7.825698244184431e-07,
"loss": 0.9029,
"step": 482
},
{
"epoch": 1.1867321867321867,
"grad_norm": 1.3266763848904035,
"learning_rate": 7.786098268526447e-07,
"loss": 0.8801,
"step": 483
},
{
"epoch": 1.1891891891891893,
"grad_norm": 1.7962161638699905,
"learning_rate": 7.746534766211206e-07,
"loss": 0.8578,
"step": 484
},
{
"epoch": 1.1916461916461916,
"grad_norm": 1.477358511769343,
"learning_rate": 7.7070083890351e-07,
"loss": 0.9048,
"step": 485
},
{
"epoch": 1.194103194103194,
"grad_norm": 1.4147723236609893,
"learning_rate": 7.667519788182911e-07,
"loss": 0.8898,
"step": 486
},
{
"epoch": 1.1965601965601966,
"grad_norm": 1.55983989969961,
"learning_rate": 7.628069614217057e-07,
"loss": 0.9509,
"step": 487
},
{
"epoch": 1.199017199017199,
"grad_norm": 1.4383185383531423,
"learning_rate": 7.588658517066892e-07,
"loss": 0.9418,
"step": 488
},
{
"epoch": 1.2014742014742015,
"grad_norm": 1.7624942370774765,
"learning_rate": 7.549287146017988e-07,
"loss": 0.8313,
"step": 489
},
{
"epoch": 1.203931203931204,
"grad_norm": 1.67001710532923,
"learning_rate": 7.509956149701443e-07,
"loss": 0.8549,
"step": 490
},
{
"epoch": 1.2063882063882063,
"grad_norm": 1.6268288582039303,
"learning_rate": 7.470666176083191e-07,
"loss": 0.8649,
"step": 491
},
{
"epoch": 1.2088452088452089,
"grad_norm": 1.4853773676445685,
"learning_rate": 7.431417872453339e-07,
"loss": 0.8956,
"step": 492
},
{
"epoch": 1.2113022113022114,
"grad_norm": 1.548578088799907,
"learning_rate": 7.39221188541548e-07,
"loss": 0.923,
"step": 493
},
{
"epoch": 1.2137592137592137,
"grad_norm": 1.710249233454194,
"learning_rate": 7.353048860876063e-07,
"loss": 0.9336,
"step": 494
},
{
"epoch": 1.2162162162162162,
"grad_norm": 1.5916044293544283,
"learning_rate": 7.31392944403375e-07,
"loss": 0.9304,
"step": 495
},
{
"epoch": 1.2186732186732188,
"grad_norm": 1.5308549158931122,
"learning_rate": 7.274854279368758e-07,
"loss": 0.9349,
"step": 496
},
{
"epoch": 1.221130221130221,
"grad_norm": 1.5229178156260872,
"learning_rate": 7.235824010632283e-07,
"loss": 0.9153,
"step": 497
},
{
"epoch": 1.2235872235872236,
"grad_norm": 1.462343882146908,
"learning_rate": 7.196839280835875e-07,
"loss": 0.9376,
"step": 498
},
{
"epoch": 1.2260442260442261,
"grad_norm": 1.5273674514727056,
"learning_rate": 7.157900732240827e-07,
"loss": 0.9231,
"step": 499
},
{
"epoch": 1.2285012285012284,
"grad_norm": 1.396230424270659,
"learning_rate": 7.119009006347624e-07,
"loss": 0.8868,
"step": 500
},
{
"epoch": 1.230958230958231,
"grad_norm": 1.6364859015395088,
"learning_rate": 7.080164743885363e-07,
"loss": 0.9287,
"step": 501
},
{
"epoch": 1.2334152334152333,
"grad_norm": 1.5048783187743708,
"learning_rate": 7.041368584801186e-07,
"loss": 0.932,
"step": 502
},
{
"epoch": 1.2358722358722358,
"grad_norm": 1.5929336784240413,
"learning_rate": 7.002621168249758e-07,
"loss": 0.9302,
"step": 503
},
{
"epoch": 1.2383292383292384,
"grad_norm": 1.5383384042486663,
"learning_rate": 6.963923132582715e-07,
"loss": 0.9059,
"step": 504
},
{
"epoch": 1.2407862407862407,
"grad_norm": 1.6790422515254697,
"learning_rate": 6.925275115338167e-07,
"loss": 0.8862,
"step": 505
},
{
"epoch": 1.2432432432432432,
"grad_norm": 1.4423572903428208,
"learning_rate": 6.886677753230183e-07,
"loss": 0.9004,
"step": 506
},
{
"epoch": 1.2457002457002457,
"grad_norm": 1.5146851789864428,
"learning_rate": 6.848131682138303e-07,
"loss": 0.8873,
"step": 507
},
{
"epoch": 1.248157248157248,
"grad_norm": 1.7060116939979457,
"learning_rate": 6.809637537097062e-07,
"loss": 0.9047,
"step": 508
},
{
"epoch": 1.2506142506142506,
"grad_norm": 1.3733517118208705,
"learning_rate": 6.77119595228554e-07,
"loss": 0.8742,
"step": 509
},
{
"epoch": 1.253071253071253,
"grad_norm": 1.4283683639472458,
"learning_rate": 6.732807561016885e-07,
"loss": 0.8697,
"step": 510
},
{
"epoch": 1.2555282555282554,
"grad_norm": 1.4832890039194944,
"learning_rate": 6.694472995727914e-07,
"loss": 0.925,
"step": 511
},
{
"epoch": 1.257985257985258,
"grad_norm": 1.4760580473983451,
"learning_rate": 6.656192887968674e-07,
"loss": 0.8959,
"step": 512
},
{
"epoch": 1.2604422604422605,
"grad_norm": 1.4517094927145753,
"learning_rate": 6.617967868392036e-07,
"loss": 0.8935,
"step": 513
},
{
"epoch": 1.2628992628992628,
"grad_norm": 1.5125319950260128,
"learning_rate": 6.579798566743313e-07,
"loss": 0.8556,
"step": 514
},
{
"epoch": 1.2653562653562653,
"grad_norm": 1.4510038145639286,
"learning_rate": 6.541685611849886e-07,
"loss": 0.9068,
"step": 515
},
{
"epoch": 1.2678132678132679,
"grad_norm": 1.786431321409276,
"learning_rate": 6.503629631610836e-07,
"loss": 0.8951,
"step": 516
},
{
"epoch": 1.2702702702702702,
"grad_norm": 4.646846867306252,
"learning_rate": 6.465631252986608e-07,
"loss": 0.8717,
"step": 517
},
{
"epoch": 1.2727272727272727,
"grad_norm": 1.6325223465886023,
"learning_rate": 6.427691101988672e-07,
"loss": 0.9277,
"step": 518
},
{
"epoch": 1.2751842751842752,
"grad_norm": 1.46996914036538,
"learning_rate": 6.389809803669225e-07,
"loss": 0.8614,
"step": 519
},
{
"epoch": 1.2776412776412776,
"grad_norm": 1.4901572767377598,
"learning_rate": 6.351987982110879e-07,
"loss": 0.8791,
"step": 520
},
{
"epoch": 1.28009828009828,
"grad_norm": 1.3956563694145843,
"learning_rate": 6.314226260416383e-07,
"loss": 0.8682,
"step": 521
},
{
"epoch": 1.2825552825552826,
"grad_norm": 1.527881875257898,
"learning_rate": 6.276525260698363e-07,
"loss": 0.9153,
"step": 522
},
{
"epoch": 1.285012285012285,
"grad_norm": 1.5192904446038689,
"learning_rate": 6.238885604069074e-07,
"loss": 0.9263,
"step": 523
},
{
"epoch": 1.2874692874692875,
"grad_norm": 1.5953405324354368,
"learning_rate": 6.201307910630145e-07,
"loss": 0.8788,
"step": 524
},
{
"epoch": 1.28992628992629,
"grad_norm": 1.4403593350780215,
"learning_rate": 6.163792799462402e-07,
"loss": 0.9414,
"step": 525
},
{
"epoch": 1.2923832923832923,
"grad_norm": 1.4506987352971705,
"learning_rate": 6.126340888615642e-07,
"loss": 0.9307,
"step": 526
},
{
"epoch": 1.2948402948402948,
"grad_norm": 1.610182690312099,
"learning_rate": 6.088952795098441e-07,
"loss": 0.9168,
"step": 527
},
{
"epoch": 1.2972972972972974,
"grad_norm": 1.5978713160514366,
"learning_rate": 6.051629134868019e-07,
"loss": 0.8393,
"step": 528
},
{
"epoch": 1.2997542997542997,
"grad_norm": 1.3547671161607415,
"learning_rate": 6.014370522820085e-07,
"loss": 0.8972,
"step": 529
},
{
"epoch": 1.3022113022113022,
"grad_norm": 1.3090769200596557,
"learning_rate": 5.977177572778677e-07,
"loss": 0.8823,
"step": 530
},
{
"epoch": 1.3046683046683047,
"grad_norm": 1.4761807072270576,
"learning_rate": 5.940050897486089e-07,
"loss": 0.9733,
"step": 531
},
{
"epoch": 1.307125307125307,
"grad_norm": 1.5243972285729563,
"learning_rate": 5.902991108592754e-07,
"loss": 0.8859,
"step": 532
},
{
"epoch": 1.3095823095823096,
"grad_norm": 1.764985461344906,
"learning_rate": 5.865998816647171e-07,
"loss": 0.8868,
"step": 533
},
{
"epoch": 1.3120393120393121,
"grad_norm": 1.6317180984567103,
"learning_rate": 5.829074631085852e-07,
"loss": 0.9361,
"step": 534
},
{
"epoch": 1.3144963144963144,
"grad_norm": 1.4981668702031359,
"learning_rate": 5.792219160223267e-07,
"loss": 0.8472,
"step": 535
},
{
"epoch": 1.316953316953317,
"grad_norm": 2.637743576229916,
"learning_rate": 5.755433011241851e-07,
"loss": 0.8736,
"step": 536
},
{
"epoch": 1.3194103194103195,
"grad_norm": 1.3474590182260624,
"learning_rate": 5.718716790181965e-07,
"loss": 0.8794,
"step": 537
},
{
"epoch": 1.3218673218673218,
"grad_norm": 1.42787784398219,
"learning_rate": 5.682071101931936e-07,
"loss": 0.8738,
"step": 538
},
{
"epoch": 1.3243243243243243,
"grad_norm": 1.6067570011968997,
"learning_rate": 5.645496550218089e-07,
"loss": 0.8879,
"step": 539
},
{
"epoch": 1.3267813267813269,
"grad_norm": 1.5914444312640619,
"learning_rate": 5.60899373759479e-07,
"loss": 0.8754,
"step": 540
},
{
"epoch": 1.3292383292383292,
"grad_norm": 1.4758896244599573,
"learning_rate": 5.572563265434526e-07,
"loss": 0.9307,
"step": 541
},
{
"epoch": 1.3316953316953317,
"grad_norm": 1.2626534960911335,
"learning_rate": 5.536205733918007e-07,
"loss": 0.843,
"step": 542
},
{
"epoch": 1.3341523341523343,
"grad_norm": 1.701842268857865,
"learning_rate": 5.499921742024257e-07,
"loss": 0.9247,
"step": 543
},
{
"epoch": 1.3366093366093366,
"grad_norm": 1.4205425380782735,
"learning_rate": 5.463711887520759e-07,
"loss": 0.8628,
"step": 544
},
{
"epoch": 1.339066339066339,
"grad_norm": 1.9140145931124193,
"learning_rate": 5.427576766953614e-07,
"loss": 0.9073,
"step": 545
},
{
"epoch": 1.3415233415233416,
"grad_norm": 1.7824661674136197,
"learning_rate": 5.391516975637699e-07,
"loss": 0.9133,
"step": 546
},
{
"epoch": 1.343980343980344,
"grad_norm": 1.5617612083029389,
"learning_rate": 5.355533107646858e-07,
"loss": 0.9531,
"step": 547
},
{
"epoch": 1.3464373464373465,
"grad_norm": 1.5245643140200376,
"learning_rate": 5.319625755804138e-07,
"loss": 0.8915,
"step": 548
},
{
"epoch": 1.348894348894349,
"grad_norm": 1.7286911090374188,
"learning_rate": 5.283795511671994e-07,
"loss": 0.863,
"step": 549
},
{
"epoch": 1.3513513513513513,
"grad_norm": 1.6181660068040473,
"learning_rate": 5.248042965542559e-07,
"loss": 0.8852,
"step": 550
},
{
"epoch": 1.3538083538083538,
"grad_norm": 1.5075998723303305,
"learning_rate": 5.212368706427912e-07,
"loss": 0.8965,
"step": 551
},
{
"epoch": 1.3562653562653564,
"grad_norm": 1.5100541752180852,
"learning_rate": 5.176773322050381e-07,
"loss": 0.8829,
"step": 552
},
{
"epoch": 1.3587223587223587,
"grad_norm": 1.863955084942053,
"learning_rate": 5.141257398832862e-07,
"loss": 0.9475,
"step": 553
},
{
"epoch": 1.3611793611793612,
"grad_norm": 1.314918195852255,
"learning_rate": 5.105821521889147e-07,
"loss": 0.8643,
"step": 554
},
{
"epoch": 1.3636363636363638,
"grad_norm": 1.8350990255330366,
"learning_rate": 5.070466275014287e-07,
"loss": 0.9441,
"step": 555
},
{
"epoch": 1.366093366093366,
"grad_norm": 1.5322707436835765,
"learning_rate": 5.03519224067499e-07,
"loss": 0.9033,
"step": 556
},
{
"epoch": 1.3685503685503686,
"grad_norm": 1.5974243335350318,
"learning_rate": 5.000000000000002e-07,
"loss": 0.8914,
"step": 557
},
{
"epoch": 1.3710073710073711,
"grad_norm": 1.4435566226659817,
"learning_rate": 4.964890132770543e-07,
"loss": 0.878,
"step": 558
},
{
"epoch": 1.3734643734643734,
"grad_norm": 1.8477167694746348,
"learning_rate": 4.929863217410767e-07,
"loss": 0.8909,
"step": 559
},
{
"epoch": 1.375921375921376,
"grad_norm": 1.4801649083461632,
"learning_rate": 4.894919830978212e-07,
"loss": 0.9079,
"step": 560
},
{
"epoch": 1.3783783783783785,
"grad_norm": 1.6811015996059142,
"learning_rate": 4.860060549154301e-07,
"loss": 0.815,
"step": 561
},
{
"epoch": 1.3808353808353808,
"grad_norm": 2.3145109947990155,
"learning_rate": 4.825285946234874e-07,
"loss": 0.9305,
"step": 562
},
{
"epoch": 1.3832923832923834,
"grad_norm": 1.4759252526730806,
"learning_rate": 4.790596595120698e-07,
"loss": 0.8589,
"step": 563
},
{
"epoch": 1.3857493857493859,
"grad_norm": 1.3378195425536465,
"learning_rate": 4.7559930673080475e-07,
"loss": 0.8659,
"step": 564
},
{
"epoch": 1.3882063882063882,
"grad_norm": 1.6393333264623262,
"learning_rate": 4.721475932879282e-07,
"loss": 0.8618,
"step": 565
},
{
"epoch": 1.3906633906633907,
"grad_norm": 1.6916138016252256,
"learning_rate": 4.6870457604934675e-07,
"loss": 0.9028,
"step": 566
},
{
"epoch": 1.393120393120393,
"grad_norm": 1.4225654726940884,
"learning_rate": 4.6527031173769857e-07,
"loss": 0.8765,
"step": 567
},
{
"epoch": 1.3955773955773956,
"grad_norm": 1.342379943145193,
"learning_rate": 4.618448569314206e-07,
"loss": 0.8855,
"step": 568
},
{
"epoch": 1.398034398034398,
"grad_norm": 2.442614220259753,
"learning_rate": 4.584282680638154e-07,
"loss": 0.8804,
"step": 569
},
{
"epoch": 1.4004914004914004,
"grad_norm": 1.5267874267053692,
"learning_rate": 4.550206014221232e-07,
"loss": 0.9668,
"step": 570
},
{
"epoch": 1.402948402948403,
"grad_norm": 1.8692131154426705,
"learning_rate": 4.5162191314659194e-07,
"loss": 0.8979,
"step": 571
},
{
"epoch": 1.4054054054054055,
"grad_norm": 2.022804110507779,
"learning_rate": 4.48232259229554e-07,
"loss": 0.8396,
"step": 572
},
{
"epoch": 1.4078624078624078,
"grad_norm": 1.444700494646088,
"learning_rate": 4.448516955145047e-07,
"loss": 0.8798,
"step": 573
},
{
"epoch": 1.4103194103194103,
"grad_norm": 1.5147425793812193,
"learning_rate": 4.414802776951798e-07,
"loss": 0.884,
"step": 574
},
{
"epoch": 1.4127764127764126,
"grad_norm": 1.4905275337775719,
"learning_rate": 4.3811806131463946e-07,
"loss": 0.8755,
"step": 575
},
{
"epoch": 1.4152334152334152,
"grad_norm": 1.5203731499114503,
"learning_rate": 4.347651017643539e-07,
"loss": 0.9089,
"step": 576
},
{
"epoch": 1.4176904176904177,
"grad_norm": 1.4792627181832774,
"learning_rate": 4.314214542832888e-07,
"loss": 0.8995,
"step": 577
},
{
"epoch": 1.42014742014742,
"grad_norm": 1.498952396682308,
"learning_rate": 4.280871739569971e-07,
"loss": 0.9,
"step": 578
},
{
"epoch": 1.4226044226044225,
"grad_norm": 1.5993242263021854,
"learning_rate": 4.247623157167102e-07,
"loss": 0.9004,
"step": 579
},
{
"epoch": 1.425061425061425,
"grad_norm": 1.2909684524904117,
"learning_rate": 4.214469343384346e-07,
"loss": 0.9442,
"step": 580
},
{
"epoch": 1.4275184275184274,
"grad_norm": 1.6601643291459118,
"learning_rate": 4.181410844420473e-07,
"loss": 0.8924,
"step": 581
},
{
"epoch": 1.42997542997543,
"grad_norm": 1.5456219866657583,
"learning_rate": 4.148448204903977e-07,
"loss": 0.8785,
"step": 582
},
{
"epoch": 1.4324324324324325,
"grad_norm": 2.3413636182110906,
"learning_rate": 4.115581967884093e-07,
"loss": 0.8867,
"step": 583
},
{
"epoch": 1.4348894348894348,
"grad_norm": 1.600188568539889,
"learning_rate": 4.082812674821865e-07,
"loss": 0.9163,
"step": 584
},
{
"epoch": 1.4373464373464373,
"grad_norm": 1.775885867839352,
"learning_rate": 4.0501408655812044e-07,
"loss": 0.94,
"step": 585
},
{
"epoch": 1.4398034398034398,
"grad_norm": 1.4853461496779028,
"learning_rate": 4.0175670784200066e-07,
"loss": 0.8828,
"step": 586
},
{
"epoch": 1.4422604422604421,
"grad_norm": 1.3437874918142632,
"learning_rate": 3.985091849981297e-07,
"loss": 0.906,
"step": 587
},
{
"epoch": 1.4447174447174447,
"grad_norm": 1.8022847542011342,
"learning_rate": 3.952715715284363e-07,
"loss": 0.9078,
"step": 588
},
{
"epoch": 1.4471744471744472,
"grad_norm": 1.4544558952772775,
"learning_rate": 3.9204392077159544e-07,
"loss": 0.9007,
"step": 589
},
{
"epoch": 1.4496314496314495,
"grad_norm": 1.5362982163739145,
"learning_rate": 3.888262859021507e-07,
"loss": 0.9329,
"step": 590
},
{
"epoch": 1.452088452088452,
"grad_norm": 1.35776640743268,
"learning_rate": 3.856187199296358e-07,
"loss": 0.9142,
"step": 591
},
{
"epoch": 1.4545454545454546,
"grad_norm": 1.552024766985016,
"learning_rate": 3.8242127569770265e-07,
"loss": 0.8624,
"step": 592
},
{
"epoch": 1.457002457002457,
"grad_norm": 1.6837813084688251,
"learning_rate": 3.792340058832515e-07,
"loss": 0.8515,
"step": 593
},
{
"epoch": 1.4594594594594594,
"grad_norm": 1.6200568897908438,
"learning_rate": 3.760569629955613e-07,
"loss": 0.9182,
"step": 594
},
{
"epoch": 1.461916461916462,
"grad_norm": 1.3613611600691327,
"learning_rate": 3.72890199375426e-07,
"loss": 0.9212,
"step": 595
},
{
"epoch": 1.4643734643734643,
"grad_norm": 1.5082364375002426,
"learning_rate": 3.697337671942913e-07,
"loss": 0.9084,
"step": 596
},
{
"epoch": 1.4668304668304668,
"grad_norm": 1.6876950054566262,
"learning_rate": 3.6658771845339676e-07,
"loss": 0.9065,
"step": 597
},
{
"epoch": 1.4692874692874693,
"grad_norm": 1.5733845918989662,
"learning_rate": 3.634521049829169e-07,
"loss": 0.8815,
"step": 598
},
{
"epoch": 1.4717444717444716,
"grad_norm": 1.8147232098247035,
"learning_rate": 3.603269784411089e-07,
"loss": 0.9133,
"step": 599
},
{
"epoch": 1.4742014742014742,
"grad_norm": 1.4595323005431404,
"learning_rate": 3.5721239031346063e-07,
"loss": 0.8071,
"step": 600
},
{
"epoch": 1.4766584766584767,
"grad_norm": 2.769727304993137,
"learning_rate": 3.541083919118438e-07,
"loss": 0.9423,
"step": 601
},
{
"epoch": 1.479115479115479,
"grad_norm": 1.610693658075675,
"learning_rate": 3.5101503437366675e-07,
"loss": 0.8946,
"step": 602
},
{
"epoch": 1.4815724815724816,
"grad_norm": 1.6365761390735136,
"learning_rate": 3.479323686610329e-07,
"loss": 0.9079,
"step": 603
},
{
"epoch": 1.484029484029484,
"grad_norm": 1.5083166362769518,
"learning_rate": 3.448604455599021e-07,
"loss": 0.8975,
"step": 604
},
{
"epoch": 1.4864864864864864,
"grad_norm": 1.60995686456512,
"learning_rate": 3.4179931567925214e-07,
"loss": 0.9498,
"step": 605
},
{
"epoch": 1.488943488943489,
"grad_norm": 1.7185280239304535,
"learning_rate": 3.387490294502456e-07,
"loss": 0.906,
"step": 606
},
{
"epoch": 1.4914004914004915,
"grad_norm": 1.321610762107921,
"learning_rate": 3.357096371254008e-07,
"loss": 0.8499,
"step": 607
},
{
"epoch": 1.4938574938574938,
"grad_norm": 1.2808013142695502,
"learning_rate": 3.3268118877776064e-07,
"loss": 0.8546,
"step": 608
},
{
"epoch": 1.4963144963144963,
"grad_norm": 1.3917333892709016,
"learning_rate": 3.296637343000704e-07,
"loss": 0.8776,
"step": 609
},
{
"epoch": 1.4987714987714988,
"grad_norm": 1.9760211104870224,
"learning_rate": 3.266573234039541e-07,
"loss": 0.8829,
"step": 610
},
{
"epoch": 1.5012285012285012,
"grad_norm": 1.9349300228421435,
"learning_rate": 3.2366200561909717e-07,
"loss": 0.9721,
"step": 611
},
{
"epoch": 1.5036855036855037,
"grad_norm": 1.5825092236415208,
"learning_rate": 3.2067783029242865e-07,
"loss": 0.8569,
"step": 612
},
{
"epoch": 1.5061425061425062,
"grad_norm": 1.7071421937295232,
"learning_rate": 3.1770484658730887e-07,
"loss": 0.9586,
"step": 613
},
{
"epoch": 1.5085995085995085,
"grad_norm": 1.3114653023022718,
"learning_rate": 3.147431034827208e-07,
"loss": 0.8879,
"step": 614
},
{
"epoch": 1.511056511056511,
"grad_norm": 1.3660758453242297,
"learning_rate": 3.11792649772461e-07,
"loss": 0.9227,
"step": 615
},
{
"epoch": 1.5135135135135136,
"grad_norm": 1.5443322628991898,
"learning_rate": 3.08853534064337e-07,
"loss": 0.8756,
"step": 616
},
{
"epoch": 1.515970515970516,
"grad_norm": 1.6205632640066696,
"learning_rate": 3.0592580477936604e-07,
"loss": 0.8631,
"step": 617
},
{
"epoch": 1.5184275184275184,
"grad_norm": 1.6750050731142352,
"learning_rate": 3.0300951015097863e-07,
"loss": 0.9308,
"step": 618
},
{
"epoch": 1.520884520884521,
"grad_norm": 1.7541208454820925,
"learning_rate": 3.001046982242216e-07,
"loss": 0.9097,
"step": 619
},
{
"epoch": 1.5233415233415233,
"grad_norm": 1.9317379819011606,
"learning_rate": 2.972114168549682e-07,
"loss": 0.9212,
"step": 620
},
{
"epoch": 1.5257985257985258,
"grad_norm": 1.2650696951039682,
"learning_rate": 2.9432971370913e-07,
"loss": 0.8836,
"step": 621
},
{
"epoch": 1.5282555282555284,
"grad_norm": 1.4960196905274168,
"learning_rate": 2.914596362618701e-07,
"loss": 0.8946,
"step": 622
},
{
"epoch": 1.5307125307125307,
"grad_norm": 1.507394873293028,
"learning_rate": 2.8860123179682236e-07,
"loss": 0.9141,
"step": 623
},
{
"epoch": 1.5331695331695332,
"grad_norm": 1.3335466599194608,
"learning_rate": 2.8575454740531115e-07,
"loss": 0.8667,
"step": 624
},
{
"epoch": 1.5356265356265357,
"grad_norm": 1.522699880941422,
"learning_rate": 2.829196299855775e-07,
"loss": 0.9901,
"step": 625
},
{
"epoch": 1.538083538083538,
"grad_norm": 1.4978237170833106,
"learning_rate": 2.800965262420043e-07,
"loss": 0.8367,
"step": 626
},
{
"epoch": 1.5405405405405406,
"grad_norm": 1.386657482706535,
"learning_rate": 2.7728528268434757e-07,
"loss": 0.8411,
"step": 627
},
{
"epoch": 1.542997542997543,
"grad_norm": 1.6884446408530738,
"learning_rate": 2.744859456269718e-07,
"loss": 0.942,
"step": 628
},
{
"epoch": 1.5454545454545454,
"grad_norm": 1.5125614736127029,
"learning_rate": 2.716985611880841e-07,
"loss": 0.8738,
"step": 629
},
{
"epoch": 1.547911547911548,
"grad_norm": 1.469704160523891,
"learning_rate": 2.689231752889768e-07,
"loss": 0.9165,
"step": 630
},
{
"epoch": 1.5503685503685505,
"grad_norm": 1.4870015368902652,
"learning_rate": 2.661598336532692e-07,
"loss": 0.8944,
"step": 631
},
{
"epoch": 1.5528255528255528,
"grad_norm": 1.4322815674497984,
"learning_rate": 2.6340858180615645e-07,
"loss": 0.8968,
"step": 632
},
{
"epoch": 1.5552825552825553,
"grad_norm": 1.3623692296259504,
"learning_rate": 2.6066946507365696e-07,
"loss": 0.8935,
"step": 633
},
{
"epoch": 1.5577395577395579,
"grad_norm": 1.484336106115407,
"learning_rate": 2.5794252858186706e-07,
"loss": 0.8993,
"step": 634
},
{
"epoch": 1.5601965601965602,
"grad_norm": 1.4403746721888513,
"learning_rate": 2.552278172562181e-07,
"loss": 0.8937,
"step": 635
},
{
"epoch": 1.5626535626535627,
"grad_norm": 1.8086294010381407,
"learning_rate": 2.525253758207353e-07,
"loss": 0.939,
"step": 636
},
{
"epoch": 1.5651105651105652,
"grad_norm": 1.4766926006894956,
"learning_rate": 2.4983524879730066e-07,
"loss": 0.9023,
"step": 637
},
{
"epoch": 1.5675675675675675,
"grad_norm": 1.3776357069750294,
"learning_rate": 2.471574805049206e-07,
"loss": 0.9233,
"step": 638
},
{
"epoch": 1.57002457002457,
"grad_norm": 1.6402577176804487,
"learning_rate": 2.444921150589957e-07,
"loss": 0.9319,
"step": 639
},
{
"epoch": 1.5724815724815726,
"grad_norm": 1.633491894306784,
"learning_rate": 2.418391963705927e-07,
"loss": 0.8818,
"step": 640
},
{
"epoch": 1.574938574938575,
"grad_norm": 1.5059260563661474,
"learning_rate": 2.391987681457219e-07,
"loss": 0.9138,
"step": 641
},
{
"epoch": 1.5773955773955772,
"grad_norm": 1.474101178051224,
"learning_rate": 2.3657087388461815e-07,
"loss": 0.8847,
"step": 642
},
{
"epoch": 1.57985257985258,
"grad_norm": 1.5606071866155247,
"learning_rate": 2.339555568810221e-07,
"loss": 0.9451,
"step": 643
},
{
"epoch": 1.5823095823095823,
"grad_norm": 1.5959144837669506,
"learning_rate": 2.3135286022146782e-07,
"loss": 0.9202,
"step": 644
},
{
"epoch": 1.5847665847665846,
"grad_norm": 1.4739665598722158,
"learning_rate": 2.2876282678457438e-07,
"loss": 0.8617,
"step": 645
},
{
"epoch": 1.5872235872235874,
"grad_norm": 1.324486221068965,
"learning_rate": 2.2618549924033704e-07,
"loss": 0.8702,
"step": 646
},
{
"epoch": 1.5896805896805897,
"grad_norm": 1.3762365282216713,
"learning_rate": 2.236209200494258e-07,
"loss": 0.9137,
"step": 647
},
{
"epoch": 1.592137592137592,
"grad_norm": 1.5081639071286088,
"learning_rate": 2.210691314624854e-07,
"loss": 0.8869,
"step": 648
},
{
"epoch": 1.5945945945945947,
"grad_norm": 1.5818213999124708,
"learning_rate": 2.1853017551943987e-07,
"loss": 0.9312,
"step": 649
},
{
"epoch": 1.597051597051597,
"grad_norm": 1.7757893297397498,
"learning_rate": 2.160040940487987e-07,
"loss": 0.8902,
"step": 650
},
{
"epoch": 1.5995085995085994,
"grad_norm": 1.3784061956740852,
"learning_rate": 2.1349092866696906e-07,
"loss": 0.8984,
"step": 651
},
{
"epoch": 1.6019656019656021,
"grad_norm": 1.4232738419586703,
"learning_rate": 2.1099072077756863e-07,
"loss": 0.8676,
"step": 652
},
{
"epoch": 1.6044226044226044,
"grad_norm": 3.799736709775996,
"learning_rate": 2.0850351157074597e-07,
"loss": 0.9107,
"step": 653
},
{
"epoch": 1.6068796068796067,
"grad_norm": 2.8003735792292055,
"learning_rate": 2.060293420224991e-07,
"loss": 0.8836,
"step": 654
},
{
"epoch": 1.6093366093366095,
"grad_norm": 1.320688739165864,
"learning_rate": 2.0356825289400183e-07,
"loss": 0.8912,
"step": 655
},
{
"epoch": 1.6117936117936118,
"grad_norm": 1.9634181459213618,
"learning_rate": 2.011202847309329e-07,
"loss": 0.9238,
"step": 656
},
{
"epoch": 1.6142506142506141,
"grad_norm": 1.5624806988822288,
"learning_rate": 1.9868547786280665e-07,
"loss": 0.8707,
"step": 657
},
{
"epoch": 1.6167076167076169,
"grad_norm": 1.4107049261598619,
"learning_rate": 1.9626387240230891e-07,
"loss": 0.868,
"step": 658
},
{
"epoch": 1.6191646191646192,
"grad_norm": 1.492505632847286,
"learning_rate": 1.9385550824463725e-07,
"loss": 0.8894,
"step": 659
},
{
"epoch": 1.6216216216216215,
"grad_norm": 1.4382690869482055,
"learning_rate": 1.9146042506684223e-07,
"loss": 0.8978,
"step": 660
},
{
"epoch": 1.6240786240786242,
"grad_norm": 1.627212055293953,
"learning_rate": 1.890786623271743e-07,
"loss": 0.871,
"step": 661
},
{
"epoch": 1.6265356265356266,
"grad_norm": 1.6843094898216577,
"learning_rate": 1.8671025926443463e-07,
"loss": 0.8952,
"step": 662
},
{
"epoch": 1.6289926289926289,
"grad_norm": 1.6654696363977666,
"learning_rate": 1.8435525489732718e-07,
"loss": 0.8591,
"step": 663
},
{
"epoch": 1.6314496314496314,
"grad_norm": 1.5188779880666268,
"learning_rate": 1.8201368802381657e-07,
"loss": 0.9323,
"step": 664
},
{
"epoch": 1.633906633906634,
"grad_norm": 1.3399754864662434,
"learning_rate": 1.7968559722048903e-07,
"loss": 0.8543,
"step": 665
},
{
"epoch": 1.6363636363636362,
"grad_norm": 1.3715522224138303,
"learning_rate": 1.7737102084191703e-07,
"loss": 0.918,
"step": 666
},
{
"epoch": 1.6388206388206388,
"grad_norm": 1.6259531621746015,
"learning_rate": 1.7506999702002678e-07,
"loss": 0.8694,
"step": 667
},
{
"epoch": 1.6412776412776413,
"grad_norm": 2.365813450212486,
"learning_rate": 1.7278256366347032e-07,
"loss": 0.8874,
"step": 668
},
{
"epoch": 1.6437346437346436,
"grad_norm": 1.7862394193970157,
"learning_rate": 1.7050875845700064e-07,
"loss": 0.8803,
"step": 669
},
{
"epoch": 1.6461916461916462,
"grad_norm": 1.557725205467593,
"learning_rate": 1.682486188608523e-07,
"loss": 0.8842,
"step": 670
},
{
"epoch": 1.6486486486486487,
"grad_norm": 1.6097866192991963,
"learning_rate": 1.6600218211012218e-07,
"loss": 0.9428,
"step": 671
},
{
"epoch": 1.651105651105651,
"grad_norm": 1.4644367524635458,
"learning_rate": 1.6376948521415711e-07,
"loss": 0.9066,
"step": 672
},
{
"epoch": 1.6535626535626535,
"grad_norm": 1.6645946570893102,
"learning_rate": 1.6155056495594465e-07,
"loss": 0.905,
"step": 673
},
{
"epoch": 1.656019656019656,
"grad_norm": 1.3397486338053894,
"learning_rate": 1.5934545789150622e-07,
"loss": 0.8785,
"step": 674
},
{
"epoch": 1.6584766584766584,
"grad_norm": 1.3685689789923863,
"learning_rate": 1.5715420034929473e-07,
"loss": 0.9179,
"step": 675
},
{
"epoch": 1.660933660933661,
"grad_norm": 1.5445547208764443,
"learning_rate": 1.5497682842959747e-07,
"loss": 0.9187,
"step": 676
},
{
"epoch": 1.6633906633906634,
"grad_norm": 1.7915250554930906,
"learning_rate": 1.5281337800393968e-07,
"loss": 0.8996,
"step": 677
},
{
"epoch": 1.6658476658476657,
"grad_norm": 1.425206424154818,
"learning_rate": 1.5066388471449453e-07,
"loss": 0.9244,
"step": 678
},
{
"epoch": 1.6683046683046683,
"grad_norm": 1.3711937617299463,
"learning_rate": 1.4852838397349542e-07,
"loss": 0.9239,
"step": 679
},
{
"epoch": 1.6707616707616708,
"grad_norm": 1.4206125252604587,
"learning_rate": 1.4640691096265357e-07,
"loss": 0.9127,
"step": 680
},
{
"epoch": 1.6732186732186731,
"grad_norm": 1.5824939121397916,
"learning_rate": 1.4429950063257712e-07,
"loss": 0.9243,
"step": 681
},
{
"epoch": 1.6756756756756757,
"grad_norm": 1.5523209337418313,
"learning_rate": 1.4220618770219604e-07,
"loss": 0.8743,
"step": 682
},
{
"epoch": 1.6781326781326782,
"grad_norm": 2.449668270011937,
"learning_rate": 1.4012700665818988e-07,
"loss": 0.9022,
"step": 683
},
{
"epoch": 1.6805896805896805,
"grad_norm": 2.797795087084986,
"learning_rate": 1.3806199175442046e-07,
"loss": 0.9127,
"step": 684
},
{
"epoch": 1.683046683046683,
"grad_norm": 1.594299579779607,
"learning_rate": 1.3601117701136588e-07,
"loss": 0.8958,
"step": 685
},
{
"epoch": 1.6855036855036856,
"grad_norm": 1.3744841196118298,
"learning_rate": 1.3397459621556128e-07,
"loss": 0.8847,
"step": 686
},
{
"epoch": 1.6879606879606879,
"grad_norm": 1.4967014666487075,
"learning_rate": 1.319522829190427e-07,
"loss": 0.9375,
"step": 687
},
{
"epoch": 1.6904176904176904,
"grad_norm": 1.257226409390929,
"learning_rate": 1.2994427043879275e-07,
"loss": 0.87,
"step": 688
},
{
"epoch": 1.692874692874693,
"grad_norm": 1.5489462493350459,
"learning_rate": 1.279505918561923e-07,
"loss": 0.8997,
"step": 689
},
{
"epoch": 1.6953316953316953,
"grad_norm": 1.478167106209186,
"learning_rate": 1.2597128001647695e-07,
"loss": 0.88,
"step": 690
},
{
"epoch": 1.6977886977886978,
"grad_norm": 1.530749437315356,
"learning_rate": 1.2400636752819337e-07,
"loss": 0.9155,
"step": 691
},
{
"epoch": 1.7002457002457003,
"grad_norm": 1.312749085771371,
"learning_rate": 1.2205588676266386e-07,
"loss": 0.9148,
"step": 692
},
{
"epoch": 1.7027027027027026,
"grad_norm": 1.2944478484019626,
"learning_rate": 1.2011986985345313e-07,
"loss": 0.9421,
"step": 693
},
{
"epoch": 1.7051597051597052,
"grad_norm": 1.9652193892846588,
"learning_rate": 1.1819834869583767e-07,
"loss": 0.8827,
"step": 694
},
{
"epoch": 1.7076167076167077,
"grad_norm": 1.4459180494396398,
"learning_rate": 1.1629135494628095e-07,
"loss": 0.9373,
"step": 695
},
{
"epoch": 1.71007371007371,
"grad_norm": 1.4835720268158477,
"learning_rate": 1.1439892002191231e-07,
"loss": 0.8919,
"step": 696
},
{
"epoch": 1.7125307125307125,
"grad_norm": 1.8195889649889634,
"learning_rate": 1.1252107510000841e-07,
"loss": 0.8234,
"step": 697
},
{
"epoch": 1.714987714987715,
"grad_norm": 1.3885033939390314,
"learning_rate": 1.1065785111748117e-07,
"loss": 0.8727,
"step": 698
},
{
"epoch": 1.7174447174447174,
"grad_norm": 1.8847414217621874,
"learning_rate": 1.0880927877036606e-07,
"loss": 0.9219,
"step": 699
},
{
"epoch": 1.71990171990172,
"grad_norm": 1.4570728152695678,
"learning_rate": 1.0697538851331767e-07,
"loss": 0.8921,
"step": 700
},
{
"epoch": 1.7223587223587224,
"grad_norm": 1.6297111465619418,
"learning_rate": 1.0515621055910817e-07,
"loss": 0.8523,
"step": 701
},
{
"epoch": 1.7248157248157248,
"grad_norm": 1.3305859076255666,
"learning_rate": 1.0335177487812862e-07,
"loss": 0.8873,
"step": 702
},
{
"epoch": 1.7272727272727273,
"grad_norm": 1.4781364488325777,
"learning_rate": 1.0156211119789581e-07,
"loss": 0.9512,
"step": 703
},
{
"epoch": 1.7297297297297298,
"grad_norm": 1.3860356549035335,
"learning_rate": 9.978724900256264e-08,
"loss": 0.816,
"step": 704
},
{
"epoch": 1.7321867321867321,
"grad_norm": 1.8635382664585312,
"learning_rate": 9.802721753243193e-08,
"loss": 0.8981,
"step": 705
},
{
"epoch": 1.7346437346437347,
"grad_norm": 1.7250194554756548,
"learning_rate": 9.628204578347466e-08,
"loss": 0.8479,
"step": 706
},
{
"epoch": 1.7371007371007372,
"grad_norm": 1.697724825854203,
"learning_rate": 9.455176250685337e-08,
"loss": 0.8986,
"step": 707
},
{
"epoch": 1.7395577395577395,
"grad_norm": 1.6181977387870101,
"learning_rate": 9.283639620844686e-08,
"loss": 0.8641,
"step": 708
},
{
"epoch": 1.742014742014742,
"grad_norm": 1.318237926606282,
"learning_rate": 9.113597514838134e-08,
"loss": 0.9309,
"step": 709
},
{
"epoch": 1.7444717444717446,
"grad_norm": 2.254843209174964,
"learning_rate": 8.94505273405658e-08,
"loss": 0.94,
"step": 710
},
{
"epoch": 1.746928746928747,
"grad_norm": 1.362243307286928,
"learning_rate": 8.778008055222807e-08,
"loss": 0.8757,
"step": 711
},
{
"epoch": 1.7493857493857494,
"grad_norm": 1.6314398152119172,
"learning_rate": 8.612466230346016e-08,
"loss": 0.8814,
"step": 712
},
{
"epoch": 1.751842751842752,
"grad_norm": 1.754218537191555,
"learning_rate": 8.448429986676298e-08,
"loss": 0.9166,
"step": 713
},
{
"epoch": 1.7542997542997543,
"grad_norm": 1.4911185442079653,
"learning_rate": 8.285902026659741e-08,
"loss": 0.9213,
"step": 714
},
{
"epoch": 1.7567567567567568,
"grad_norm": 1.4620109120308156,
"learning_rate": 8.124885027894013e-08,
"loss": 0.9285,
"step": 715
},
{
"epoch": 1.7592137592137593,
"grad_norm": 1.5197712931486254,
"learning_rate": 7.965381643084068e-08,
"loss": 0.8578,
"step": 716
},
{
"epoch": 1.7616707616707616,
"grad_norm": 1.3554180543190764,
"learning_rate": 7.807394499998576e-08,
"loss": 0.8945,
"step": 717
},
{
"epoch": 1.7641277641277642,
"grad_norm": 1.7074271145927862,
"learning_rate": 7.650926201426634e-08,
"loss": 0.9014,
"step": 718
},
{
"epoch": 1.7665847665847667,
"grad_norm": 1.7305821717380618,
"learning_rate": 7.495979325134805e-08,
"loss": 0.9427,
"step": 719
},
{
"epoch": 1.769041769041769,
"grad_norm": 1.6549796883627756,
"learning_rate": 7.342556423824675e-08,
"loss": 0.8929,
"step": 720
},
{
"epoch": 1.7714987714987716,
"grad_norm": 1.693588021942915,
"learning_rate": 7.19066002509089e-08,
"loss": 0.9149,
"step": 721
},
{
"epoch": 1.773955773955774,
"grad_norm": 1.3883677751773678,
"learning_rate": 7.040292631379386e-08,
"loss": 0.9379,
"step": 722
},
{
"epoch": 1.7764127764127764,
"grad_norm": 1.415233971009828,
"learning_rate": 6.891456719946187e-08,
"loss": 0.8786,
"step": 723
},
{
"epoch": 1.7788697788697787,
"grad_norm": 1.3887984611992066,
"learning_rate": 6.744154742816721e-08,
"loss": 0.8809,
"step": 724
},
{
"epoch": 1.7813267813267815,
"grad_norm": 1.4599586632095534,
"learning_rate": 6.598389126745207e-08,
"loss": 0.8906,
"step": 725
},
{
"epoch": 1.7837837837837838,
"grad_norm": 1.3618867212694985,
"learning_rate": 6.454162273174879e-08,
"loss": 0.8875,
"step": 726
},
{
"epoch": 1.786240786240786,
"grad_norm": 1.616703409825558,
"learning_rate": 6.311476558198337e-08,
"loss": 0.9186,
"step": 727
},
{
"epoch": 1.7886977886977888,
"grad_norm": 1.5643574404054077,
"learning_rate": 6.170334332518323e-08,
"loss": 0.8191,
"step": 728
},
{
"epoch": 1.7911547911547911,
"grad_norm": 1.6547417394218757,
"learning_rate": 6.030737921409168e-08,
"loss": 0.9015,
"step": 729
},
{
"epoch": 1.7936117936117935,
"grad_norm": 1.3258614962104087,
"learning_rate": 5.892689624678326e-08,
"loss": 0.9003,
"step": 730
},
{
"epoch": 1.7960687960687962,
"grad_norm": 1.4886792779962632,
"learning_rate": 5.756191716628556e-08,
"loss": 0.9197,
"step": 731
},
{
"epoch": 1.7985257985257985,
"grad_norm": 1.4228121623738594,
"learning_rate": 5.621246446020489e-08,
"loss": 0.9151,
"step": 732
},
{
"epoch": 1.8009828009828008,
"grad_norm": 1.4694783098206843,
"learning_rate": 5.487856036035488e-08,
"loss": 0.9478,
"step": 733
},
{
"epoch": 1.8034398034398036,
"grad_norm": 1.6835863076974373,
"learning_rate": 5.356022684239059e-08,
"loss": 0.8852,
"step": 734
},
{
"epoch": 1.805896805896806,
"grad_norm": 1.468678481999939,
"learning_rate": 5.22574856254474e-08,
"loss": 0.9227,
"step": 735
},
{
"epoch": 1.8083538083538082,
"grad_norm": 1.664873807912385,
"learning_rate": 5.0970358171781525e-08,
"loss": 0.9709,
"step": 736
},
{
"epoch": 1.810810810810811,
"grad_norm": 1.4869146515566096,
"learning_rate": 4.969886568641757e-08,
"loss": 0.8627,
"step": 737
},
{
"epoch": 1.8132678132678133,
"grad_norm": 1.5302504751076207,
"learning_rate": 4.844302911679909e-08,
"loss": 0.8694,
"step": 738
},
{
"epoch": 1.8157248157248156,
"grad_norm": 1.3703447874730004,
"learning_rate": 4.720286915244309e-08,
"loss": 0.8388,
"step": 739
},
{
"epoch": 1.8181818181818183,
"grad_norm": 1.8909215620115458,
"learning_rate": 4.597840622459936e-08,
"loss": 0.8697,
"step": 740
},
{
"epoch": 1.8206388206388207,
"grad_norm": 1.8396393900066417,
"learning_rate": 4.476966050591413e-08,
"loss": 0.9238,
"step": 741
},
{
"epoch": 1.823095823095823,
"grad_norm": 1.4266543724530107,
"learning_rate": 4.3576651910097076e-08,
"loss": 0.8901,
"step": 742
},
{
"epoch": 1.8255528255528255,
"grad_norm": 1.3873776301248721,
"learning_rate": 4.239940009159415e-08,
"loss": 0.904,
"step": 743
},
{
"epoch": 1.828009828009828,
"grad_norm": 1.554746782909149,
"learning_rate": 4.1237924445262864e-08,
"loss": 0.8605,
"step": 744
},
{
"epoch": 1.8304668304668303,
"grad_norm": 1.5101859068421868,
"learning_rate": 4.0092244106053386e-08,
"loss": 0.8981,
"step": 745
},
{
"epoch": 1.8329238329238329,
"grad_norm": 1.559988660277713,
"learning_rate": 3.896237794869339e-08,
"loss": 0.8617,
"step": 746
},
{
"epoch": 1.8353808353808354,
"grad_norm": 1.7088654168690152,
"learning_rate": 3.7848344587376294e-08,
"loss": 0.8696,
"step": 747
},
{
"epoch": 1.8378378378378377,
"grad_norm": 1.6151695281653022,
"learning_rate": 3.675016237545536e-08,
"loss": 0.9057,
"step": 748
},
{
"epoch": 1.8402948402948403,
"grad_norm": 1.6260854298806076,
"learning_rate": 3.566784940514145e-08,
"loss": 0.8715,
"step": 749
},
{
"epoch": 1.8427518427518428,
"grad_norm": 1.5483874580947579,
"learning_rate": 3.460142350720396e-08,
"loss": 0.8928,
"step": 750
},
{
"epoch": 1.845208845208845,
"grad_norm": 1.3856178647115844,
"learning_rate": 3.3550902250677914e-08,
"loss": 0.8898,
"step": 751
},
{
"epoch": 1.8476658476658476,
"grad_norm": 1.6217388770422025,
"learning_rate": 3.251630294257479e-08,
"loss": 0.8887,
"step": 752
},
{
"epoch": 1.8501228501228502,
"grad_norm": 1.7461700477068736,
"learning_rate": 3.1497642627596245e-08,
"loss": 0.9144,
"step": 753
},
{
"epoch": 1.8525798525798525,
"grad_norm": 1.3204466111990083,
"learning_rate": 3.0494938087854306e-08,
"loss": 0.9046,
"step": 754
},
{
"epoch": 1.855036855036855,
"grad_norm": 1.5311966806895796,
"learning_rate": 2.9508205842594724e-08,
"loss": 0.9614,
"step": 755
},
{
"epoch": 1.8574938574938575,
"grad_norm": 1.5618296845089266,
"learning_rate": 2.8537462147924184e-08,
"loss": 0.9499,
"step": 756
},
{
"epoch": 1.8599508599508598,
"grad_norm": 1.506551170833026,
"learning_rate": 2.7582722996543405e-08,
"loss": 0.8963,
"step": 757
},
{
"epoch": 1.8624078624078624,
"grad_norm": 1.6976856223987622,
"learning_rate": 2.6644004117483355e-08,
"loss": 0.938,
"step": 758
},
{
"epoch": 1.864864864864865,
"grad_norm": 1.7152873001996989,
"learning_rate": 2.572132097584523e-08,
"loss": 0.9302,
"step": 759
},
{
"epoch": 1.8673218673218672,
"grad_norm": 1.4747780984909165,
"learning_rate": 2.481468877254722e-08,
"loss": 0.8927,
"step": 760
},
{
"epoch": 1.8697788697788698,
"grad_norm": 1.6455964184060006,
"learning_rate": 2.3924122444072936e-08,
"loss": 0.92,
"step": 761
},
{
"epoch": 1.8722358722358723,
"grad_norm": 1.4135913992521028,
"learning_rate": 2.3049636662225702e-08,
"loss": 0.9095,
"step": 762
},
{
"epoch": 1.8746928746928746,
"grad_norm": 1.3283327210815499,
"learning_rate": 2.2191245833886985e-08,
"loss": 0.923,
"step": 763
},
{
"epoch": 1.8771498771498771,
"grad_norm": 2.1206540317643316,
"learning_rate": 2.134896410077891e-08,
"loss": 0.8759,
"step": 764
},
{
"epoch": 1.8796068796068797,
"grad_norm": 1.4686301829717907,
"learning_rate": 2.052280533923101e-08,
"loss": 0.91,
"step": 765
},
{
"epoch": 1.882063882063882,
"grad_norm": 2.2833412802518835,
"learning_rate": 1.9712783159952063e-08,
"loss": 0.8681,
"step": 766
},
{
"epoch": 1.8845208845208845,
"grad_norm": 1.5664189962805568,
"learning_rate": 1.8918910907805728e-08,
"loss": 0.8958,
"step": 767
},
{
"epoch": 1.886977886977887,
"grad_norm": 1.3597898472460088,
"learning_rate": 1.81412016615905e-08,
"loss": 0.8846,
"step": 768
},
{
"epoch": 1.8894348894348894,
"grad_norm": 1.794391013663922,
"learning_rate": 1.737966823382442e-08,
"loss": 0.8704,
"step": 769
},
{
"epoch": 1.8918918918918919,
"grad_norm": 1.3876663428773923,
"learning_rate": 1.6634323170533926e-08,
"loss": 0.8826,
"step": 770
},
{
"epoch": 1.8943488943488944,
"grad_norm": 1.964992745747424,
"learning_rate": 1.5905178751047134e-08,
"loss": 0.9233,
"step": 771
},
{
"epoch": 1.8968058968058967,
"grad_norm": 1.5474022069405389,
"learning_rate": 1.519224698779198e-08,
"loss": 0.9574,
"step": 772
},
{
"epoch": 1.8992628992628993,
"grad_norm": 1.367308795693324,
"learning_rate": 1.4495539626097287e-08,
"loss": 0.8783,
"step": 773
},
{
"epoch": 1.9017199017199018,
"grad_norm": 1.4446894160381547,
"learning_rate": 1.3815068144000352e-08,
"loss": 0.9278,
"step": 774
},
{
"epoch": 1.904176904176904,
"grad_norm": 1.4909159417858395,
"learning_rate": 1.3150843752057438e-08,
"loss": 0.8681,
"step": 775
},
{
"epoch": 1.9066339066339066,
"grad_norm": 1.358589936876708,
"learning_rate": 1.2502877393158585e-08,
"loss": 0.8662,
"step": 776
},
{
"epoch": 1.9090909090909092,
"grad_norm": 1.4371079365419295,
"learning_rate": 1.1871179742348413e-08,
"loss": 0.8844,
"step": 777
},
{
"epoch": 1.9115479115479115,
"grad_norm": 1.5058059815383584,
"learning_rate": 1.1255761206649061e-08,
"loss": 0.8885,
"step": 778
},
{
"epoch": 1.914004914004914,
"grad_norm": 1.4300739849238244,
"learning_rate": 1.0656631924889748e-08,
"loss": 0.8997,
"step": 779
},
{
"epoch": 1.9164619164619165,
"grad_norm": 1.5658510974559385,
"learning_rate": 1.0073801767539247e-08,
"loss": 0.8983,
"step": 780
},
{
"epoch": 1.9189189189189189,
"grad_norm": 1.4965427543973944,
"learning_rate": 9.507280336542911e-09,
"loss": 0.8676,
"step": 781
},
{
"epoch": 1.9213759213759214,
"grad_norm": 1.36581135793563,
"learning_rate": 8.957076965165234e-09,
"loss": 0.8792,
"step": 782
},
{
"epoch": 1.923832923832924,
"grad_norm": 1.5156402618603095,
"learning_rate": 8.423200717835976e-09,
"loss": 0.8626,
"step": 783
},
{
"epoch": 1.9262899262899262,
"grad_norm": 1.3074563333752685,
"learning_rate": 7.905660390000069e-09,
"loss": 0.9315,
"step": 784
},
{
"epoch": 1.9287469287469288,
"grad_norm": 1.7609221141327198,
"learning_rate": 7.404464507973607e-09,
"loss": 0.8862,
"step": 785
},
{
"epoch": 1.9312039312039313,
"grad_norm": 1.5426817574716647,
"learning_rate": 6.919621328802972e-09,
"loss": 0.9543,
"step": 786
},
{
"epoch": 1.9336609336609336,
"grad_norm": 2.3064282757005663,
"learning_rate": 6.4511388401286e-09,
"loss": 0.9583,
"step": 787
},
{
"epoch": 1.9361179361179361,
"grad_norm": 1.67978354162131,
"learning_rate": 5.999024760054094e-09,
"loss": 0.9364,
"step": 788
},
{
"epoch": 1.9385749385749387,
"grad_norm": 2.269285024101559,
"learning_rate": 5.563286537018319e-09,
"loss": 0.858,
"step": 789
},
{
"epoch": 1.941031941031941,
"grad_norm": 1.7349413650625205,
"learning_rate": 5.143931349672837e-09,
"loss": 0.9019,
"step": 790
},
{
"epoch": 1.9434889434889435,
"grad_norm": 1.652477600280537,
"learning_rate": 4.7409661067642215e-09,
"loss": 0.9325,
"step": 791
},
{
"epoch": 1.945945945945946,
"grad_norm": 1.354819301610776,
"learning_rate": 4.354397447019375e-09,
"loss": 0.9047,
"step": 792
},
{
"epoch": 1.9484029484029484,
"grad_norm": 1.615707852796759,
"learning_rate": 3.9842317390369474e-09,
"loss": 0.9033,
"step": 793
},
{
"epoch": 1.950859950859951,
"grad_norm": 1.4784764684056093,
"learning_rate": 3.630475081181861e-09,
"loss": 0.8419,
"step": 794
},
{
"epoch": 1.9533169533169534,
"grad_norm": 1.316270670020737,
"learning_rate": 3.2931333014850626e-09,
"loss": 0.8478,
"step": 795
},
{
"epoch": 1.9557739557739557,
"grad_norm": 1.4172812525176979,
"learning_rate": 2.972211957547377e-09,
"loss": 0.8208,
"step": 796
},
{
"epoch": 1.9582309582309583,
"grad_norm": 1.6364175608068232,
"learning_rate": 2.667716336448356e-09,
"loss": 0.9062,
"step": 797
},
{
"epoch": 1.9606879606879608,
"grad_norm": 1.6122614160580477,
"learning_rate": 2.379651454658571e-09,
"loss": 0.8939,
"step": 798
},
{
"epoch": 1.9631449631449631,
"grad_norm": 1.5173118240679087,
"learning_rate": 2.108022057957348e-09,
"loss": 0.8899,
"step": 799
},
{
"epoch": 1.9656019656019657,
"grad_norm": 1.4032962894445904,
"learning_rate": 1.8528326213548273e-09,
"loss": 0.9269,
"step": 800
},
{
"epoch": 1.9680589680589682,
"grad_norm": 1.491232169954221,
"learning_rate": 1.6140873490172457e-09,
"loss": 0.887,
"step": 801
},
{
"epoch": 1.9705159705159705,
"grad_norm": 1.452706323624227,
"learning_rate": 1.3917901741989923e-09,
"loss": 0.9033,
"step": 802
},
{
"epoch": 1.972972972972973,
"grad_norm": 2.9039670137491176,
"learning_rate": 1.1859447591769934e-09,
"loss": 0.8912,
"step": 803
},
{
"epoch": 1.9754299754299756,
"grad_norm": 1.426897627539779,
"learning_rate": 9.96554495190205e-10,
"loss": 0.8626,
"step": 804
},
{
"epoch": 1.9778869778869779,
"grad_norm": 1.333240805775527,
"learning_rate": 8.236225023844357e-10,
"loss": 0.8773,
"step": 805
},
{
"epoch": 1.9803439803439802,
"grad_norm": 1.3907079742810688,
"learning_rate": 6.671516297606094e-10,
"loss": 0.9252,
"step": 806
},
{
"epoch": 1.982800982800983,
"grad_norm": 1.698786893046404,
"learning_rate": 5.271444551276927e-10,
"loss": 0.9504,
"step": 807
},
{
"epoch": 1.9852579852579852,
"grad_norm": 1.8549325727891992,
"learning_rate": 4.0360328506017226e-10,
"loss": 0.8953,
"step": 808
},
{
"epoch": 1.9877149877149876,
"grad_norm": 1.4404540052198962,
"learning_rate": 2.965301548606414e-10,
"loss": 0.8946,
"step": 809
},
{
"epoch": 1.9901719901719903,
"grad_norm": 1.2420916620768656,
"learning_rate": 2.0592682852549338e-10,
"loss": 0.8543,
"step": 810
},
{
"epoch": 1.9926289926289926,
"grad_norm": 1.5902830913290869,
"learning_rate": 1.317947987163892e-10,
"loss": 0.9287,
"step": 811
},
{
"epoch": 1.995085995085995,
"grad_norm": 1.7356553731894724,
"learning_rate": 7.41352867354994e-11,
"loss": 0.9318,
"step": 812
},
{
"epoch": 1.9975429975429977,
"grad_norm": 1.3632994827873626,
"learning_rate": 3.294924250540898e-11,
"loss": 0.9022,
"step": 813
},
{
"epoch": 2.0,
"grad_norm": 1.5450655192747615,
"learning_rate": 8.237344553241321e-12,
"loss": 0.9407,
"step": 814
}
],
"logging_steps": 1,
"max_steps": 814,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 407,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 8.602630032143155e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}