{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.98971898560658, "eval_steps": 500, "global_step": 910, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0054832076764907475, "grad_norm": 604.3109811491834, "learning_rate": 8.791208791208792e-07, "loss": 3.4565, "step": 1 }, { "epoch": 0.010966415352981495, "grad_norm": 602.1736348957847, "learning_rate": 1.7582417582417585e-06, "loss": 3.445, "step": 2 }, { "epoch": 0.01644962302947224, "grad_norm": 572.1582926794415, "learning_rate": 2.6373626373626375e-06, "loss": 3.426, "step": 3 }, { "epoch": 0.02193283070596299, "grad_norm": 301.9672730610099, "learning_rate": 3.516483516483517e-06, "loss": 3.2554, "step": 4 }, { "epoch": 0.027416038382453736, "grad_norm": 157.4890153555457, "learning_rate": 4.395604395604396e-06, "loss": 2.9123, "step": 5 }, { "epoch": 0.03289924605894448, "grad_norm": 93.35477967830393, "learning_rate": 5.274725274725275e-06, "loss": 2.6655, "step": 6 }, { "epoch": 0.03838245373543523, "grad_norm": 23.960002925273315, "learning_rate": 6.153846153846155e-06, "loss": 2.526, "step": 7 }, { "epoch": 0.04386566141192598, "grad_norm": 16.617925535315912, "learning_rate": 7.032967032967034e-06, "loss": 2.3244, "step": 8 }, { "epoch": 0.04934886908841672, "grad_norm": 13.65445534016627, "learning_rate": 7.912087912087913e-06, "loss": 2.2167, "step": 9 }, { "epoch": 0.05483207676490747, "grad_norm": 23.60213319182947, "learning_rate": 8.791208791208792e-06, "loss": 2.1214, "step": 10 }, { "epoch": 0.06031528444139822, "grad_norm": 9.750741286283983, "learning_rate": 9.670329670329671e-06, "loss": 2.0631, "step": 11 }, { "epoch": 0.06579849211788896, "grad_norm": 13.69478328276252, "learning_rate": 1.054945054945055e-05, "loss": 2.0035, "step": 12 }, { "epoch": 0.07128169979437972, "grad_norm": 8.863133767549522, "learning_rate": 1.1428571428571429e-05, "loss": 1.9018, "step": 13 }, { "epoch": 0.07676490747087046, "grad_norm": 6.184273756096322, "learning_rate": 1.230769230769231e-05, "loss": 1.8409, "step": 14 }, { "epoch": 0.0822481151473612, "grad_norm": 8.463726907785716, "learning_rate": 1.3186813186813187e-05, "loss": 1.7862, "step": 15 }, { "epoch": 0.08773132282385196, "grad_norm": 5.9606779223424455, "learning_rate": 1.4065934065934068e-05, "loss": 1.7449, "step": 16 }, { "epoch": 0.0932145305003427, "grad_norm": 18.546012018929744, "learning_rate": 1.4945054945054947e-05, "loss": 1.7071, "step": 17 }, { "epoch": 0.09869773817683344, "grad_norm": 9.23610239091234, "learning_rate": 1.5824175824175826e-05, "loss": 1.7254, "step": 18 }, { "epoch": 0.1041809458533242, "grad_norm": 44.47081850643636, "learning_rate": 1.6703296703296707e-05, "loss": 1.6708, "step": 19 }, { "epoch": 0.10966415352981494, "grad_norm": 29.350706185079904, "learning_rate": 1.7582417582417584e-05, "loss": 1.7401, "step": 20 }, { "epoch": 0.11514736120630568, "grad_norm": 7.168404792591547, "learning_rate": 1.8461538461538465e-05, "loss": 1.6881, "step": 21 }, { "epoch": 0.12063056888279644, "grad_norm": 28.75909337885957, "learning_rate": 1.9340659340659342e-05, "loss": 1.5999, "step": 22 }, { "epoch": 0.12611377655928718, "grad_norm": 12.270722231807452, "learning_rate": 2.021978021978022e-05, "loss": 1.7174, "step": 23 }, { "epoch": 0.13159698423577793, "grad_norm": 51.182199832000634, "learning_rate": 2.10989010989011e-05, "loss": 1.6164, "step": 24 }, { "epoch": 0.13708019191226867, "grad_norm": 5.983294276939527, "learning_rate": 2.197802197802198e-05, "loss": 1.6279, "step": 25 }, { "epoch": 0.14256339958875944, "grad_norm": 12.36006380527371, "learning_rate": 2.2857142857142858e-05, "loss": 1.5335, "step": 26 }, { "epoch": 0.14804660726525018, "grad_norm": 3.912477135556689, "learning_rate": 2.373626373626374e-05, "loss": 1.5637, "step": 27 }, { "epoch": 0.15352981494174092, "grad_norm": 8.798850308211108, "learning_rate": 2.461538461538462e-05, "loss": 1.546, "step": 28 }, { "epoch": 0.15901302261823166, "grad_norm": 6.133566343370423, "learning_rate": 2.5494505494505493e-05, "loss": 1.5493, "step": 29 }, { "epoch": 0.1644962302947224, "grad_norm": 3.0393729351601753, "learning_rate": 2.6373626373626374e-05, "loss": 1.4988, "step": 30 }, { "epoch": 0.16997943797121315, "grad_norm": 28.575328415461772, "learning_rate": 2.7252747252747255e-05, "loss": 1.5026, "step": 31 }, { "epoch": 0.17546264564770392, "grad_norm": 27.101171796307156, "learning_rate": 2.8131868131868136e-05, "loss": 1.6522, "step": 32 }, { "epoch": 0.18094585332419466, "grad_norm": 7.286329527124996, "learning_rate": 2.9010989010989013e-05, "loss": 1.6616, "step": 33 }, { "epoch": 0.1864290610006854, "grad_norm": 4.786509337900977, "learning_rate": 2.9890109890109894e-05, "loss": 1.495, "step": 34 }, { "epoch": 0.19191226867717615, "grad_norm": 2.804143471750863, "learning_rate": 3.0769230769230774e-05, "loss": 1.4495, "step": 35 }, { "epoch": 0.1973954763536669, "grad_norm": 9.304401113296121, "learning_rate": 3.164835164835165e-05, "loss": 1.4403, "step": 36 }, { "epoch": 0.20287868403015763, "grad_norm": 5.457009814887685, "learning_rate": 3.252747252747253e-05, "loss": 1.4336, "step": 37 }, { "epoch": 0.2083618917066484, "grad_norm": 3.7957565393477437, "learning_rate": 3.340659340659341e-05, "loss": 1.4243, "step": 38 }, { "epoch": 0.21384509938313914, "grad_norm": 4.9220861710974075, "learning_rate": 3.4285714285714284e-05, "loss": 1.3984, "step": 39 }, { "epoch": 0.21932830705962988, "grad_norm": 4.736412206172175, "learning_rate": 3.516483516483517e-05, "loss": 1.3668, "step": 40 }, { "epoch": 0.22481151473612063, "grad_norm": 5.581594123475227, "learning_rate": 3.6043956043956045e-05, "loss": 1.3513, "step": 41 }, { "epoch": 0.23029472241261137, "grad_norm": 3.5585171099348427, "learning_rate": 3.692307692307693e-05, "loss": 1.3557, "step": 42 }, { "epoch": 0.2357779300891021, "grad_norm": 5.217567824661255, "learning_rate": 3.7802197802197807e-05, "loss": 1.3368, "step": 43 }, { "epoch": 0.24126113776559288, "grad_norm": 3.8941164699166784, "learning_rate": 3.8681318681318684e-05, "loss": 1.3192, "step": 44 }, { "epoch": 0.24674434544208362, "grad_norm": 3.3715823234615345, "learning_rate": 3.956043956043957e-05, "loss": 1.3194, "step": 45 }, { "epoch": 0.25222755311857437, "grad_norm": 4.071045479574351, "learning_rate": 4.043956043956044e-05, "loss": 1.3, "step": 46 }, { "epoch": 0.2577107607950651, "grad_norm": 4.790439556324944, "learning_rate": 4.131868131868133e-05, "loss": 1.3051, "step": 47 }, { "epoch": 0.26319396847155585, "grad_norm": 3.6760783714364793, "learning_rate": 4.21978021978022e-05, "loss": 1.2885, "step": 48 }, { "epoch": 0.2686771761480466, "grad_norm": 3.19497049785778, "learning_rate": 4.307692307692308e-05, "loss": 1.2962, "step": 49 }, { "epoch": 0.27416038382453733, "grad_norm": 2.620333577151409, "learning_rate": 4.395604395604396e-05, "loss": 1.2452, "step": 50 }, { "epoch": 0.2796435915010281, "grad_norm": 3.573140694736915, "learning_rate": 4.483516483516484e-05, "loss": 1.2912, "step": 51 }, { "epoch": 0.2851267991775189, "grad_norm": 2.3723854257821513, "learning_rate": 4.5714285714285716e-05, "loss": 1.2468, "step": 52 }, { "epoch": 0.2906100068540096, "grad_norm": 2.5628938638391277, "learning_rate": 4.65934065934066e-05, "loss": 1.2883, "step": 53 }, { "epoch": 0.29609321453050036, "grad_norm": 3.2477927500052184, "learning_rate": 4.747252747252748e-05, "loss": 1.2864, "step": 54 }, { "epoch": 0.3015764222069911, "grad_norm": 2.4110262583540716, "learning_rate": 4.8351648351648355e-05, "loss": 1.2886, "step": 55 }, { "epoch": 0.30705962988348184, "grad_norm": 3.9054257622250614, "learning_rate": 4.923076923076924e-05, "loss": 1.2544, "step": 56 }, { "epoch": 0.3125428375599726, "grad_norm": 2.684900989864461, "learning_rate": 5.0109890109890116e-05, "loss": 1.2506, "step": 57 }, { "epoch": 0.31802604523646333, "grad_norm": 3.1294292463297673, "learning_rate": 5.098901098901099e-05, "loss": 1.2619, "step": 58 }, { "epoch": 0.32350925291295407, "grad_norm": 4.302163409683541, "learning_rate": 5.186813186813188e-05, "loss": 1.2531, "step": 59 }, { "epoch": 0.3289924605894448, "grad_norm": 2.3198593094902566, "learning_rate": 5.274725274725275e-05, "loss": 1.2444, "step": 60 }, { "epoch": 0.33447566826593556, "grad_norm": 3.075723813655852, "learning_rate": 5.3626373626373626e-05, "loss": 1.2318, "step": 61 }, { "epoch": 0.3399588759424263, "grad_norm": 2.751318001428128, "learning_rate": 5.450549450549451e-05, "loss": 1.2283, "step": 62 }, { "epoch": 0.34544208361891704, "grad_norm": 3.792517967186425, "learning_rate": 5.538461538461539e-05, "loss": 1.2347, "step": 63 }, { "epoch": 0.35092529129540784, "grad_norm": 2.502677006785022, "learning_rate": 5.626373626373627e-05, "loss": 1.2154, "step": 64 }, { "epoch": 0.3564084989718986, "grad_norm": 2.4447089296161955, "learning_rate": 5.714285714285715e-05, "loss": 1.2062, "step": 65 }, { "epoch": 0.3618917066483893, "grad_norm": 2.7612204760550183, "learning_rate": 5.8021978021978026e-05, "loss": 1.2105, "step": 66 }, { "epoch": 0.36737491432488006, "grad_norm": 3.3730773809799968, "learning_rate": 5.890109890109891e-05, "loss": 1.2129, "step": 67 }, { "epoch": 0.3728581220013708, "grad_norm": 2.067462539301751, "learning_rate": 5.978021978021979e-05, "loss": 1.2021, "step": 68 }, { "epoch": 0.37834132967786155, "grad_norm": 3.367733224494582, "learning_rate": 6.0659340659340665e-05, "loss": 1.2207, "step": 69 }, { "epoch": 0.3838245373543523, "grad_norm": 2.1443388252000335, "learning_rate": 6.153846153846155e-05, "loss": 1.2025, "step": 70 }, { "epoch": 0.38930774503084303, "grad_norm": 3.4158140446082994, "learning_rate": 6.241758241758242e-05, "loss": 1.2132, "step": 71 }, { "epoch": 0.3947909527073338, "grad_norm": 2.58088978525386, "learning_rate": 6.32967032967033e-05, "loss": 1.1843, "step": 72 }, { "epoch": 0.4002741603838245, "grad_norm": 2.897001887626188, "learning_rate": 6.417582417582419e-05, "loss": 1.1996, "step": 73 }, { "epoch": 0.40575736806031526, "grad_norm": 3.0910986197243573, "learning_rate": 6.505494505494506e-05, "loss": 1.215, "step": 74 }, { "epoch": 0.41124057573680606, "grad_norm": 2.36902708485275, "learning_rate": 6.593406593406594e-05, "loss": 1.172, "step": 75 }, { "epoch": 0.4167237834132968, "grad_norm": 2.9090585544359366, "learning_rate": 6.681318681318683e-05, "loss": 1.192, "step": 76 }, { "epoch": 0.42220699108978754, "grad_norm": 2.682374090006958, "learning_rate": 6.76923076923077e-05, "loss": 1.1929, "step": 77 }, { "epoch": 0.4276901987662783, "grad_norm": 2.6472629365834037, "learning_rate": 6.857142857142857e-05, "loss": 1.1578, "step": 78 }, { "epoch": 0.433173406442769, "grad_norm": 2.58566062639449, "learning_rate": 6.945054945054945e-05, "loss": 1.153, "step": 79 }, { "epoch": 0.43865661411925977, "grad_norm": 2.7406033885199848, "learning_rate": 7.032967032967034e-05, "loss": 1.1697, "step": 80 }, { "epoch": 0.4441398217957505, "grad_norm": 2.137577114283556, "learning_rate": 7.12087912087912e-05, "loss": 1.1492, "step": 81 }, { "epoch": 0.44962302947224125, "grad_norm": 2.9855017798950976, "learning_rate": 7.208791208791209e-05, "loss": 1.1584, "step": 82 }, { "epoch": 0.455106237148732, "grad_norm": 2.220583427738456, "learning_rate": 7.296703296703297e-05, "loss": 1.1437, "step": 83 }, { "epoch": 0.46058944482522274, "grad_norm": 2.706030331520761, "learning_rate": 7.384615384615386e-05, "loss": 1.1673, "step": 84 }, { "epoch": 0.4660726525017135, "grad_norm": 3.190829729449872, "learning_rate": 7.472527472527473e-05, "loss": 1.1413, "step": 85 }, { "epoch": 0.4715558601782042, "grad_norm": 1.9236806571413316, "learning_rate": 7.560439560439561e-05, "loss": 1.1546, "step": 86 }, { "epoch": 0.477039067854695, "grad_norm": 2.9656722488411, "learning_rate": 7.64835164835165e-05, "loss": 1.1705, "step": 87 }, { "epoch": 0.48252227553118576, "grad_norm": 1.7873438452480972, "learning_rate": 7.736263736263737e-05, "loss": 1.1512, "step": 88 }, { "epoch": 0.4880054832076765, "grad_norm": 3.366153588378385, "learning_rate": 7.824175824175825e-05, "loss": 1.1854, "step": 89 }, { "epoch": 0.49348869088416725, "grad_norm": 2.3979761080289306, "learning_rate": 7.912087912087914e-05, "loss": 1.1785, "step": 90 }, { "epoch": 0.498971898560658, "grad_norm": 2.68813176726075, "learning_rate": 8e-05, "loss": 1.1689, "step": 91 }, { "epoch": 0.5044551062371487, "grad_norm": 3.634421050069424, "learning_rate": 7.999970571955439e-05, "loss": 1.1534, "step": 92 }, { "epoch": 0.5099383139136395, "grad_norm": 2.110037855631259, "learning_rate": 7.999882288254757e-05, "loss": 1.1398, "step": 93 }, { "epoch": 0.5154215215901302, "grad_norm": 3.029801321990504, "learning_rate": 7.999735150196965e-05, "loss": 1.15, "step": 94 }, { "epoch": 0.520904729266621, "grad_norm": 1.9878631276509169, "learning_rate": 7.999529159947053e-05, "loss": 1.1524, "step": 95 }, { "epoch": 0.5263879369431117, "grad_norm": 2.997439251632094, "learning_rate": 7.999264320535968e-05, "loss": 1.161, "step": 96 }, { "epoch": 0.5318711446196025, "grad_norm": 1.7997765525760119, "learning_rate": 7.998940635860564e-05, "loss": 1.1392, "step": 97 }, { "epoch": 0.5373543522960932, "grad_norm": 2.3934821316457353, "learning_rate": 7.998558110683542e-05, "loss": 1.1511, "step": 98 }, { "epoch": 0.542837559972584, "grad_norm": 2.394286604934712, "learning_rate": 7.998116750633388e-05, "loss": 1.1482, "step": 99 }, { "epoch": 0.5483207676490747, "grad_norm": 2.5638379332850447, "learning_rate": 7.997616562204282e-05, "loss": 1.132, "step": 100 }, { "epoch": 0.5538039753255655, "grad_norm": 2.4073701211461715, "learning_rate": 7.99705755275601e-05, "loss": 1.1653, "step": 101 }, { "epoch": 0.5592871830020562, "grad_norm": 2.9908052478558127, "learning_rate": 7.996439730513846e-05, "loss": 1.1437, "step": 102 }, { "epoch": 0.564770390678547, "grad_norm": 2.338649947963062, "learning_rate": 7.995763104568444e-05, "loss": 1.1257, "step": 103 }, { "epoch": 0.5702535983550377, "grad_norm": 2.1324859396744853, "learning_rate": 7.99502768487569e-05, "loss": 1.1082, "step": 104 }, { "epoch": 0.5757368060315284, "grad_norm": 1.8207695969892468, "learning_rate": 7.994233482256567e-05, "loss": 1.1324, "step": 105 }, { "epoch": 0.5812200137080192, "grad_norm": 3.0671728842532793, "learning_rate": 7.993380508396992e-05, "loss": 1.1673, "step": 106 }, { "epoch": 0.5867032213845099, "grad_norm": 1.2132079987970015, "learning_rate": 7.992468775847638e-05, "loss": 1.1332, "step": 107 }, { "epoch": 0.5921864290610007, "grad_norm": 2.8722682327215647, "learning_rate": 7.99149829802376e-05, "loss": 1.1937, "step": 108 }, { "epoch": 0.5976696367374914, "grad_norm": 1.578063060667415, "learning_rate": 7.990469089204992e-05, "loss": 1.1516, "step": 109 }, { "epoch": 0.6031528444139822, "grad_norm": 2.4024809989588336, "learning_rate": 7.989381164535131e-05, "loss": 1.1776, "step": 110 }, { "epoch": 0.6086360520904729, "grad_norm": 1.7795121351579433, "learning_rate": 7.988234540021928e-05, "loss": 1.153, "step": 111 }, { "epoch": 0.6141192597669637, "grad_norm": 2.5507155124877032, "learning_rate": 7.987029232536841e-05, "loss": 1.1546, "step": 112 }, { "epoch": 0.6196024674434544, "grad_norm": 2.7262135259336095, "learning_rate": 7.98576525981479e-05, "loss": 1.1681, "step": 113 }, { "epoch": 0.6250856751199452, "grad_norm": 1.756472159467694, "learning_rate": 7.9844426404539e-05, "loss": 1.1337, "step": 114 }, { "epoch": 0.630568882796436, "grad_norm": 3.664337440475356, "learning_rate": 7.983061393915222e-05, "loss": 1.1458, "step": 115 }, { "epoch": 0.6360520904729267, "grad_norm": 1.951846935673031, "learning_rate": 7.981621540522444e-05, "loss": 1.1322, "step": 116 }, { "epoch": 0.6415352981494175, "grad_norm": 1.7976750357904085, "learning_rate": 7.980123101461606e-05, "loss": 1.1263, "step": 117 }, { "epoch": 0.6470185058259081, "grad_norm": 2.300277427581849, "learning_rate": 7.978566098780771e-05, "loss": 1.1413, "step": 118 }, { "epoch": 0.6525017135023989, "grad_norm": 2.360648922071168, "learning_rate": 7.976950555389713e-05, "loss": 1.1274, "step": 119 }, { "epoch": 0.6579849211788896, "grad_norm": 2.0913575786408147, "learning_rate": 7.97527649505957e-05, "loss": 1.1213, "step": 120 }, { "epoch": 0.6634681288553804, "grad_norm": 2.022249517249403, "learning_rate": 7.973543942422506e-05, "loss": 1.1175, "step": 121 }, { "epoch": 0.6689513365318711, "grad_norm": 2.2294222010405362, "learning_rate": 7.97175292297134e-05, "loss": 1.1281, "step": 122 }, { "epoch": 0.6744345442083619, "grad_norm": 1.7133244378182648, "learning_rate": 7.969903463059169e-05, "loss": 1.1099, "step": 123 }, { "epoch": 0.6799177518848526, "grad_norm": 2.0157523028333775, "learning_rate": 7.96799558989899e-05, "loss": 1.1175, "step": 124 }, { "epoch": 0.6854009595613434, "grad_norm": 1.7297097607076724, "learning_rate": 7.966029331563287e-05, "loss": 1.1131, "step": 125 }, { "epoch": 0.6908841672378341, "grad_norm": 2.2317660437246034, "learning_rate": 7.964004716983635e-05, "loss": 1.1066, "step": 126 }, { "epoch": 0.6963673749143249, "grad_norm": 1.9857286499887847, "learning_rate": 7.961921775950254e-05, "loss": 1.1083, "step": 127 }, { "epoch": 0.7018505825908157, "grad_norm": 1.6946992419248856, "learning_rate": 7.959780539111585e-05, "loss": 1.0985, "step": 128 }, { "epoch": 0.7073337902673064, "grad_norm": 1.949379922441364, "learning_rate": 7.957581037973835e-05, "loss": 1.0792, "step": 129 }, { "epoch": 0.7128169979437972, "grad_norm": 2.0374474232968454, "learning_rate": 7.955323304900514e-05, "loss": 1.1081, "step": 130 }, { "epoch": 0.7183002056202878, "grad_norm": 1.5382297067387796, "learning_rate": 7.953007373111956e-05, "loss": 1.0894, "step": 131 }, { "epoch": 0.7237834132967786, "grad_norm": 1.994423662926792, "learning_rate": 7.950633276684833e-05, "loss": 1.0971, "step": 132 }, { "epoch": 0.7292666209732693, "grad_norm": 1.6956272845939566, "learning_rate": 7.948201050551651e-05, "loss": 1.078, "step": 133 }, { "epoch": 0.7347498286497601, "grad_norm": 1.808121530050818, "learning_rate": 7.945710730500243e-05, "loss": 1.087, "step": 134 }, { "epoch": 0.7402330363262508, "grad_norm": 1.6031163127588859, "learning_rate": 7.943162353173232e-05, "loss": 1.0878, "step": 135 }, { "epoch": 0.7457162440027416, "grad_norm": 2.8069443037445057, "learning_rate": 7.940555956067495e-05, "loss": 1.0852, "step": 136 }, { "epoch": 0.7511994516792323, "grad_norm": 1.023257465419051, "learning_rate": 7.937891577533624e-05, "loss": 1.0864, "step": 137 }, { "epoch": 0.7566826593557231, "grad_norm": 3.7981840568943266, "learning_rate": 7.93516925677534e-05, "loss": 1.1577, "step": 138 }, { "epoch": 0.7621658670322139, "grad_norm": 1.957367420792266, "learning_rate": 7.932389033848931e-05, "loss": 1.1428, "step": 139 }, { "epoch": 0.7676490747087046, "grad_norm": 2.620391650383465, "learning_rate": 7.929550949662659e-05, "loss": 1.1262, "step": 140 }, { "epoch": 0.7731322823851954, "grad_norm": 1.8055917441902372, "learning_rate": 7.92665504597616e-05, "loss": 1.1087, "step": 141 }, { "epoch": 0.7786154900616861, "grad_norm": 1.5140320658198094, "learning_rate": 7.923701365399826e-05, "loss": 1.1206, "step": 142 }, { "epoch": 0.7840986977381769, "grad_norm": 2.3311158871467983, "learning_rate": 7.920689951394175e-05, "loss": 1.1354, "step": 143 }, { "epoch": 0.7895819054146676, "grad_norm": 2.871009250239358, "learning_rate": 7.917620848269224e-05, "loss": 1.1159, "step": 144 }, { "epoch": 0.7950651130911583, "grad_norm": 1.174246832683915, "learning_rate": 7.914494101183822e-05, "loss": 1.112, "step": 145 }, { "epoch": 0.800548320767649, "grad_norm": 1.9715656445082748, "learning_rate": 7.911309756144995e-05, "loss": 1.1233, "step": 146 }, { "epoch": 0.8060315284441398, "grad_norm": 1.3667047276351785, "learning_rate": 7.908067860007268e-05, "loss": 1.0966, "step": 147 }, { "epoch": 0.8115147361206305, "grad_norm": 1.8908076306855728, "learning_rate": 7.904768460471975e-05, "loss": 1.118, "step": 148 }, { "epoch": 0.8169979437971213, "grad_norm": 1.6160590718366654, "learning_rate": 7.90141160608655e-05, "loss": 1.0836, "step": 149 }, { "epoch": 0.8224811514736121, "grad_norm": 2.257542807740706, "learning_rate": 7.897997346243825e-05, "loss": 1.0985, "step": 150 }, { "epoch": 0.8279643591501028, "grad_norm": 1.2533944841043951, "learning_rate": 7.894525731181297e-05, "loss": 1.0678, "step": 151 }, { "epoch": 0.8334475668265936, "grad_norm": 2.4996679708615988, "learning_rate": 7.890996811980386e-05, "loss": 1.1143, "step": 152 }, { "epoch": 0.8389307745030843, "grad_norm": 1.6215671222990462, "learning_rate": 7.887410640565689e-05, "loss": 1.0947, "step": 153 }, { "epoch": 0.8444139821795751, "grad_norm": 1.4102568455851006, "learning_rate": 7.883767269704209e-05, "loss": 1.0635, "step": 154 }, { "epoch": 0.8498971898560658, "grad_norm": 2.0199914067223763, "learning_rate": 7.880066753004588e-05, "loss": 1.1007, "step": 155 }, { "epoch": 0.8553803975325566, "grad_norm": 1.4233619206252466, "learning_rate": 7.876309144916312e-05, "loss": 1.0964, "step": 156 }, { "epoch": 0.8608636052090473, "grad_norm": 1.5315953952088068, "learning_rate": 7.87249450072891e-05, "loss": 1.1091, "step": 157 }, { "epoch": 0.866346812885538, "grad_norm": 1.5672306259553663, "learning_rate": 7.86862287657114e-05, "loss": 1.0961, "step": 158 }, { "epoch": 0.8718300205620287, "grad_norm": 1.722337979582118, "learning_rate": 7.864694329410168e-05, "loss": 1.1032, "step": 159 }, { "epoch": 0.8773132282385195, "grad_norm": 1.1707259582896943, "learning_rate": 7.860708917050722e-05, "loss": 1.083, "step": 160 }, { "epoch": 0.8827964359150103, "grad_norm": 2.1260596185722216, "learning_rate": 7.85666669813425e-05, "loss": 1.0775, "step": 161 }, { "epoch": 0.888279643591501, "grad_norm": 1.1049918316810494, "learning_rate": 7.852567732138051e-05, "loss": 1.0804, "step": 162 }, { "epoch": 0.8937628512679918, "grad_norm": 1.5314067038496302, "learning_rate": 7.848412079374403e-05, "loss": 1.0931, "step": 163 }, { "epoch": 0.8992460589444825, "grad_norm": 1.526207291286965, "learning_rate": 7.844199800989672e-05, "loss": 1.126, "step": 164 }, { "epoch": 0.9047292666209733, "grad_norm": 1.4889589302339163, "learning_rate": 7.839930958963415e-05, "loss": 1.0688, "step": 165 }, { "epoch": 0.910212474297464, "grad_norm": 1.8224369715346593, "learning_rate": 7.835605616107471e-05, "loss": 1.0981, "step": 166 }, { "epoch": 0.9156956819739548, "grad_norm": 1.9809507356149727, "learning_rate": 7.83122383606503e-05, "loss": 1.0716, "step": 167 }, { "epoch": 0.9211788896504455, "grad_norm": 1.319128673805323, "learning_rate": 7.826785683309702e-05, "loss": 1.0511, "step": 168 }, { "epoch": 0.9266620973269363, "grad_norm": 1.7628250921107829, "learning_rate": 7.822291223144564e-05, "loss": 1.0815, "step": 169 }, { "epoch": 0.932145305003427, "grad_norm": 1.5619673215880712, "learning_rate": 7.817740521701204e-05, "loss": 1.0851, "step": 170 }, { "epoch": 0.9376285126799178, "grad_norm": 1.6877201996768514, "learning_rate": 7.813133645938744e-05, "loss": 1.0655, "step": 171 }, { "epoch": 0.9431117203564084, "grad_norm": 1.4453595064062965, "learning_rate": 7.808470663642856e-05, "loss": 1.0468, "step": 172 }, { "epoch": 0.9485949280328992, "grad_norm": 1.3988144950022152, "learning_rate": 7.803751643424769e-05, "loss": 1.0684, "step": 173 }, { "epoch": 0.95407813570939, "grad_norm": 1.829632986890131, "learning_rate": 7.798976654720248e-05, "loss": 1.0658, "step": 174 }, { "epoch": 0.9595613433858807, "grad_norm": 1.160599126459967, "learning_rate": 7.794145767788582e-05, "loss": 1.0666, "step": 175 }, { "epoch": 0.9650445510623715, "grad_norm": 1.6311368539718085, "learning_rate": 7.789259053711554e-05, "loss": 1.0732, "step": 176 }, { "epoch": 0.9705277587388622, "grad_norm": 1.3145190591456402, "learning_rate": 7.784316584392379e-05, "loss": 1.0564, "step": 177 }, { "epoch": 0.976010966415353, "grad_norm": 1.434756052069605, "learning_rate": 7.779318432554663e-05, "loss": 1.0624, "step": 178 }, { "epoch": 0.9814941740918437, "grad_norm": 1.3078999315572475, "learning_rate": 7.774264671741324e-05, "loss": 1.0486, "step": 179 }, { "epoch": 0.9869773817683345, "grad_norm": 1.4474824177372518, "learning_rate": 7.769155376313509e-05, "loss": 1.0511, "step": 180 }, { "epoch": 0.9924605894448252, "grad_norm": 1.4306506181803345, "learning_rate": 7.763990621449507e-05, "loss": 1.0657, "step": 181 }, { "epoch": 0.997943797121316, "grad_norm": 1.5854524936397603, "learning_rate": 7.758770483143634e-05, "loss": 1.0638, "step": 182 }, { "epoch": 1.0034270047978067, "grad_norm": 2.818624037527729, "learning_rate": 7.753495038205123e-05, "loss": 1.7465, "step": 183 }, { "epoch": 1.0089102124742975, "grad_norm": 1.2531970095723952, "learning_rate": 7.748164364256989e-05, "loss": 1.0425, "step": 184 }, { "epoch": 1.0143934201507883, "grad_norm": 1.8473887590213065, "learning_rate": 7.742778539734884e-05, "loss": 1.0309, "step": 185 }, { "epoch": 1.019876627827279, "grad_norm": 1.1978766650574866, "learning_rate": 7.737337643885956e-05, "loss": 1.0365, "step": 186 }, { "epoch": 1.0253598355037696, "grad_norm": 1.6327228122841784, "learning_rate": 7.73184175676766e-05, "loss": 1.0412, "step": 187 }, { "epoch": 1.0308430431802604, "grad_norm": 1.038869594675331, "learning_rate": 7.726290959246606e-05, "loss": 1.0325, "step": 188 }, { "epoch": 1.0363262508567512, "grad_norm": 1.8931974706378887, "learning_rate": 7.720685332997354e-05, "loss": 1.0575, "step": 189 }, { "epoch": 1.041809458533242, "grad_norm": 1.0099409227594682, "learning_rate": 7.715024960501209e-05, "loss": 1.0356, "step": 190 }, { "epoch": 1.0472926662097326, "grad_norm": 1.7017523773128445, "learning_rate": 7.709309925045023e-05, "loss": 1.048, "step": 191 }, { "epoch": 1.0527758738862234, "grad_norm": 1.0632647956537205, "learning_rate": 7.70354031071995e-05, "loss": 1.0619, "step": 192 }, { "epoch": 1.0582590815627142, "grad_norm": 1.6480122219648587, "learning_rate": 7.697716202420227e-05, "loss": 1.0626, "step": 193 }, { "epoch": 1.063742289239205, "grad_norm": 1.337533514643958, "learning_rate": 7.691837685841913e-05, "loss": 1.0521, "step": 194 }, { "epoch": 1.0692254969156956, "grad_norm": 1.405204005470527, "learning_rate": 7.685904847481631e-05, "loss": 1.0488, "step": 195 }, { "epoch": 1.0747087045921864, "grad_norm": 1.0789567262015418, "learning_rate": 7.679917774635298e-05, "loss": 1.0354, "step": 196 }, { "epoch": 1.0801919122686772, "grad_norm": 1.5059550846436562, "learning_rate": 7.673876555396835e-05, "loss": 1.0488, "step": 197 }, { "epoch": 1.085675119945168, "grad_norm": 1.285703383014824, "learning_rate": 7.667781278656879e-05, "loss": 1.0326, "step": 198 }, { "epoch": 1.0911583276216588, "grad_norm": 1.166732355574816, "learning_rate": 7.661632034101466e-05, "loss": 1.0322, "step": 199 }, { "epoch": 1.0966415352981493, "grad_norm": 1.6867809480621192, "learning_rate": 7.655428912210718e-05, "loss": 1.0364, "step": 200 }, { "epoch": 1.1021247429746401, "grad_norm": 1.2531266932796477, "learning_rate": 7.64917200425751e-05, "loss": 1.0374, "step": 201 }, { "epoch": 1.107607950651131, "grad_norm": 1.6183973189746483, "learning_rate": 7.642861402306123e-05, "loss": 1.0302, "step": 202 }, { "epoch": 1.1130911583276217, "grad_norm": 1.5885719958570816, "learning_rate": 7.636497199210895e-05, "loss": 1.0312, "step": 203 }, { "epoch": 1.1185743660041123, "grad_norm": 1.271489486550103, "learning_rate": 7.630079488614853e-05, "loss": 1.0439, "step": 204 }, { "epoch": 1.124057573680603, "grad_norm": 1.6077174464582225, "learning_rate": 7.623608364948334e-05, "loss": 1.0369, "step": 205 }, { "epoch": 1.129540781357094, "grad_norm": 1.0526829284737067, "learning_rate": 7.617083923427596e-05, "loss": 1.0294, "step": 206 }, { "epoch": 1.1350239890335847, "grad_norm": 1.8026775192610152, "learning_rate": 7.610506260053415e-05, "loss": 1.0249, "step": 207 }, { "epoch": 1.1405071967100753, "grad_norm": 1.1092234118793582, "learning_rate": 7.603875471609677e-05, "loss": 1.0479, "step": 208 }, { "epoch": 1.145990404386566, "grad_norm": 1.252308944775191, "learning_rate": 7.597191655661952e-05, "loss": 1.0377, "step": 209 }, { "epoch": 1.1514736120630569, "grad_norm": 1.481646296662436, "learning_rate": 7.590454910556058e-05, "loss": 1.0475, "step": 210 }, { "epoch": 1.1569568197395477, "grad_norm": 1.1007433182898057, "learning_rate": 7.583665335416608e-05, "loss": 1.0105, "step": 211 }, { "epoch": 1.1624400274160385, "grad_norm": 1.2445540839929121, "learning_rate": 7.576823030145566e-05, "loss": 1.0236, "step": 212 }, { "epoch": 1.167923235092529, "grad_norm": 1.4019899879808024, "learning_rate": 7.569928095420762e-05, "loss": 1.0437, "step": 213 }, { "epoch": 1.1734064427690198, "grad_norm": 1.293921037672153, "learning_rate": 7.562980632694421e-05, "loss": 1.0078, "step": 214 }, { "epoch": 1.1788896504455106, "grad_norm": 1.574451864443491, "learning_rate": 7.555980744191666e-05, "loss": 1.0183, "step": 215 }, { "epoch": 1.1843728581220014, "grad_norm": 1.0261703383565228, "learning_rate": 7.548928532909006e-05, "loss": 1.0243, "step": 216 }, { "epoch": 1.1898560657984922, "grad_norm": 1.3888687863743303, "learning_rate": 7.541824102612839e-05, "loss": 1.0241, "step": 217 }, { "epoch": 1.1953392734749828, "grad_norm": 1.435498169074072, "learning_rate": 7.534667557837912e-05, "loss": 1.0348, "step": 218 }, { "epoch": 1.2008224811514736, "grad_norm": 1.1812741919729888, "learning_rate": 7.527459003885783e-05, "loss": 1.013, "step": 219 }, { "epoch": 1.2063056888279644, "grad_norm": 1.209488409322031, "learning_rate": 7.520198546823275e-05, "loss": 1.0279, "step": 220 }, { "epoch": 1.211788896504455, "grad_norm": 1.5861672644314255, "learning_rate": 7.512886293480914e-05, "loss": 1.0217, "step": 221 }, { "epoch": 1.2172721041809458, "grad_norm": 1.0784984727297116, "learning_rate": 7.505522351451363e-05, "loss": 1.0308, "step": 222 }, { "epoch": 1.2227553118574366, "grad_norm": 1.4013682384096804, "learning_rate": 7.498106829087822e-05, "loss": 1.0209, "step": 223 }, { "epoch": 1.2282385195339274, "grad_norm": 1.1069329108834502, "learning_rate": 7.490639835502458e-05, "loss": 1.0184, "step": 224 }, { "epoch": 1.2337217272104182, "grad_norm": 1.2292930770165162, "learning_rate": 7.483121480564779e-05, "loss": 1.0192, "step": 225 }, { "epoch": 1.2392049348869087, "grad_norm": 1.1091280093145537, "learning_rate": 7.475551874900027e-05, "loss": 1.0227, "step": 226 }, { "epoch": 1.2446881425633995, "grad_norm": 1.3883445894297355, "learning_rate": 7.467931129887548e-05, "loss": 1.0317, "step": 227 }, { "epoch": 1.2501713502398903, "grad_norm": 1.3258593612132867, "learning_rate": 7.460259357659155e-05, "loss": 1.0276, "step": 228 }, { "epoch": 1.2556545579163811, "grad_norm": 1.2962983275547162, "learning_rate": 7.452536671097476e-05, "loss": 1.0155, "step": 229 }, { "epoch": 1.261137765592872, "grad_norm": 1.2345976366071654, "learning_rate": 7.444763183834292e-05, "loss": 0.9999, "step": 230 }, { "epoch": 1.2666209732693625, "grad_norm": 1.371675205771468, "learning_rate": 7.436939010248868e-05, "loss": 1.0242, "step": 231 }, { "epoch": 1.2721041809458533, "grad_norm": 1.043563875055389, "learning_rate": 7.429064265466269e-05, "loss": 0.9978, "step": 232 }, { "epoch": 1.2775873886223441, "grad_norm": 1.4841522976495107, "learning_rate": 7.421139065355663e-05, "loss": 1.0152, "step": 233 }, { "epoch": 1.2830705962988347, "grad_norm": 0.9605681175633184, "learning_rate": 7.413163526528623e-05, "loss": 1.0162, "step": 234 }, { "epoch": 1.2885538039753255, "grad_norm": 1.2067100095706547, "learning_rate": 7.405137766337406e-05, "loss": 1.0565, "step": 235 }, { "epoch": 1.2940370116518163, "grad_norm": 1.5489835530200247, "learning_rate": 7.397061902873223e-05, "loss": 1.0316, "step": 236 }, { "epoch": 1.299520219328307, "grad_norm": 0.936867837906044, "learning_rate": 7.388936054964512e-05, "loss": 1.0218, "step": 237 }, { "epoch": 1.3050034270047979, "grad_norm": 1.3113159040022304, "learning_rate": 7.38076034217518e-05, "loss": 1.0305, "step": 238 }, { "epoch": 1.3104866346812885, "grad_norm": 0.9712447789123708, "learning_rate": 7.372534884802844e-05, "loss": 1.0403, "step": 239 }, { "epoch": 1.3159698423577793, "grad_norm": 1.2727545703286138, "learning_rate": 7.364259803877072e-05, "loss": 1.0366, "step": 240 }, { "epoch": 1.32145305003427, "grad_norm": 1.1239260262608586, "learning_rate": 7.355935221157584e-05, "loss": 1.0031, "step": 241 }, { "epoch": 1.3269362577107608, "grad_norm": 1.1122104721303026, "learning_rate": 7.347561259132479e-05, "loss": 1.0358, "step": 242 }, { "epoch": 1.3324194653872516, "grad_norm": 1.4676521237806195, "learning_rate": 7.33913804101642e-05, "loss": 1.0362, "step": 243 }, { "epoch": 1.3379026730637422, "grad_norm": 1.2984708218697247, "learning_rate": 7.330665690748825e-05, "loss": 1.0095, "step": 244 }, { "epoch": 1.343385880740233, "grad_norm": 1.2692918018410921, "learning_rate": 7.322144332992047e-05, "loss": 1.0126, "step": 245 }, { "epoch": 1.3488690884167238, "grad_norm": 0.9963650435822852, "learning_rate": 7.313574093129532e-05, "loss": 0.9937, "step": 246 }, { "epoch": 1.3543522960932146, "grad_norm": 1.45458036698363, "learning_rate": 7.30495509726398e-05, "loss": 1.0389, "step": 247 }, { "epoch": 1.3598355037697054, "grad_norm": 1.0844293357139385, "learning_rate": 7.29628747221549e-05, "loss": 1.042, "step": 248 }, { "epoch": 1.365318711446196, "grad_norm": 1.3150162241275731, "learning_rate": 7.287571345519688e-05, "loss": 1.0076, "step": 249 }, { "epoch": 1.3708019191226868, "grad_norm": 1.170486608498072, "learning_rate": 7.278806845425856e-05, "loss": 1.0411, "step": 250 }, { "epoch": 1.3762851267991776, "grad_norm": 1.0369874340952356, "learning_rate": 7.269994100895047e-05, "loss": 1.0137, "step": 251 }, { "epoch": 1.3817683344756682, "grad_norm": 1.3592340170974726, "learning_rate": 7.261133241598177e-05, "loss": 1.0104, "step": 252 }, { "epoch": 1.387251542152159, "grad_norm": 1.0387909603628305, "learning_rate": 7.25222439791413e-05, "loss": 1.0039, "step": 253 }, { "epoch": 1.3927347498286498, "grad_norm": 1.2165525508974753, "learning_rate": 7.24326770092783e-05, "loss": 1.0117, "step": 254 }, { "epoch": 1.3982179575051406, "grad_norm": 1.0298641450176182, "learning_rate": 7.234263282428312e-05, "loss": 1.0087, "step": 255 }, { "epoch": 1.4037011651816313, "grad_norm": 2.3062469579312768, "learning_rate": 7.225211274906795e-05, "loss": 1.0376, "step": 256 }, { "epoch": 1.409184372858122, "grad_norm": 1.0448171443413676, "learning_rate": 7.216111811554718e-05, "loss": 1.015, "step": 257 }, { "epoch": 1.4146675805346127, "grad_norm": 2.282441256400205, "learning_rate": 7.206965026261787e-05, "loss": 1.0506, "step": 258 }, { "epoch": 1.4201507882111035, "grad_norm": 1.5228239109337287, "learning_rate": 7.197771053614006e-05, "loss": 1.0488, "step": 259 }, { "epoch": 1.4256339958875943, "grad_norm": 7.1797647736282535, "learning_rate": 7.188530028891691e-05, "loss": 1.0477, "step": 260 }, { "epoch": 1.4311172035640851, "grad_norm": 19.858381633007593, "learning_rate": 7.179242088067487e-05, "loss": 1.2469, "step": 261 }, { "epoch": 1.4366004112405757, "grad_norm": 12.439075720330774, "learning_rate": 7.169907367804363e-05, "loss": 1.3376, "step": 262 }, { "epoch": 1.4420836189170665, "grad_norm": 10.336294598744168, "learning_rate": 7.160526005453599e-05, "loss": 1.2434, "step": 263 }, { "epoch": 1.4475668265935573, "grad_norm": 3325.4750490493866, "learning_rate": 7.151098139052772e-05, "loss": 4.0042, "step": 264 }, { "epoch": 1.4530500342700479, "grad_norm": 131.55901739711575, "learning_rate": 7.141623907323717e-05, "loss": 1.8859, "step": 265 }, { "epoch": 1.4585332419465387, "grad_norm": 11.388495314344222, "learning_rate": 7.13210344967049e-05, "loss": 1.3033, "step": 266 }, { "epoch": 1.4640164496230295, "grad_norm": 5.045627021720784, "learning_rate": 7.122536906177318e-05, "loss": 1.1984, "step": 267 }, { "epoch": 1.4694996572995203, "grad_norm": 63.13120968720244, "learning_rate": 7.112924417606536e-05, "loss": 1.2658, "step": 268 }, { "epoch": 1.474982864976011, "grad_norm": 78.12090408575494, "learning_rate": 7.103266125396512e-05, "loss": 1.6972, "step": 269 }, { "epoch": 1.4804660726525016, "grad_norm": 40.95778397870689, "learning_rate": 7.093562171659577e-05, "loss": 1.4509, "step": 270 }, { "epoch": 1.4859492803289924, "grad_norm": 37.962763812559935, "learning_rate": 7.083812699179919e-05, "loss": 1.34, "step": 271 }, { "epoch": 1.4914324880054832, "grad_norm": 26.82750605185721, "learning_rate": 7.074017851411495e-05, "loss": 1.2961, "step": 272 }, { "epoch": 1.496915695681974, "grad_norm": 4.438206627978028, "learning_rate": 7.064177772475912e-05, "loss": 1.2112, "step": 273 }, { "epoch": 1.5023989033584648, "grad_norm": 8.269192562168978, "learning_rate": 7.054292607160313e-05, "loss": 1.1848, "step": 274 }, { "epoch": 1.5078821110349554, "grad_norm": 3.4078464376801048, "learning_rate": 7.044362500915239e-05, "loss": 1.1018, "step": 275 }, { "epoch": 1.5133653187114462, "grad_norm": 3.8991572015827836, "learning_rate": 7.034387599852494e-05, "loss": 1.0906, "step": 276 }, { "epoch": 1.518848526387937, "grad_norm": 2.2792452805163577, "learning_rate": 7.024368050742996e-05, "loss": 1.0535, "step": 277 }, { "epoch": 1.5243317340644276, "grad_norm": 7.92182168324662, "learning_rate": 7.014304001014614e-05, "loss": 1.09, "step": 278 }, { "epoch": 1.5298149417409186, "grad_norm": 4.604463247427479, "learning_rate": 7.004195598749997e-05, "loss": 1.0929, "step": 279 }, { "epoch": 1.5352981494174092, "grad_norm": 1.6269169421948018, "learning_rate": 6.994042992684406e-05, "loss": 1.0721, "step": 280 }, { "epoch": 1.5407813570939, "grad_norm": 3.1522500098672723, "learning_rate": 6.983846332203508e-05, "loss": 1.052, "step": 281 }, { "epoch": 1.5462645647703908, "grad_norm": 1.744104230884076, "learning_rate": 6.973605767341194e-05, "loss": 1.0536, "step": 282 }, { "epoch": 1.5517477724468813, "grad_norm": 1.7435972526503143, "learning_rate": 6.963321448777367e-05, "loss": 1.0562, "step": 283 }, { "epoch": 1.5572309801233721, "grad_norm": 2.0699287078075783, "learning_rate": 6.952993527835714e-05, "loss": 1.0508, "step": 284 }, { "epoch": 1.562714187799863, "grad_norm": 2.5216940307065374, "learning_rate": 6.942622156481498e-05, "loss": 1.0696, "step": 285 }, { "epoch": 1.5681973954763535, "grad_norm": 1.5542171373751579, "learning_rate": 6.932207487319305e-05, "loss": 1.0351, "step": 286 }, { "epoch": 1.5736806031528445, "grad_norm": 1.1081954487560843, "learning_rate": 6.921749673590813e-05, "loss": 1.0399, "step": 287 }, { "epoch": 1.579163810829335, "grad_norm": 1.5993253594110808, "learning_rate": 6.911248869172523e-05, "loss": 1.0364, "step": 288 }, { "epoch": 1.584647018505826, "grad_norm": 1.077571235080621, "learning_rate": 6.900705228573507e-05, "loss": 1.0142, "step": 289 }, { "epoch": 1.5901302261823167, "grad_norm": 1.2982202905714664, "learning_rate": 6.890118906933126e-05, "loss": 1.0182, "step": 290 }, { "epoch": 1.5956134338588073, "grad_norm": 1.4040374561281856, "learning_rate": 6.879490060018754e-05, "loss": 1.0215, "step": 291 }, { "epoch": 1.6010966415352983, "grad_norm": 1.6011611520244904, "learning_rate": 6.86881884422348e-05, "loss": 0.9977, "step": 292 }, { "epoch": 1.6065798492117889, "grad_norm": 1.7528977017352316, "learning_rate": 6.858105416563812e-05, "loss": 0.9914, "step": 293 }, { "epoch": 1.6120630568882797, "grad_norm": 1.0189454039968566, "learning_rate": 6.847349934677363e-05, "loss": 1.0233, "step": 294 }, { "epoch": 1.6175462645647705, "grad_norm": 1.6918231992095927, "learning_rate": 6.836552556820533e-05, "loss": 1.0325, "step": 295 }, { "epoch": 1.623029472241261, "grad_norm": 0.9564755595265833, "learning_rate": 6.82571344186618e-05, "loss": 1.0198, "step": 296 }, { "epoch": 1.6285126799177518, "grad_norm": 1.2847377343488389, "learning_rate": 6.814832749301285e-05, "loss": 1.0148, "step": 297 }, { "epoch": 1.6339958875942426, "grad_norm": 0.9873795646279533, "learning_rate": 6.803910639224598e-05, "loss": 1.0357, "step": 298 }, { "epoch": 1.6394790952707334, "grad_norm": 1.5820758914116304, "learning_rate": 6.792947272344292e-05, "loss": 1.0376, "step": 299 }, { "epoch": 1.6449623029472242, "grad_norm": 0.901645544400314, "learning_rate": 6.78194280997559e-05, "loss": 1.0506, "step": 300 }, { "epoch": 1.6504455106237148, "grad_norm": 1.4932631104259941, "learning_rate": 6.770897414038398e-05, "loss": 1.0178, "step": 301 }, { "epoch": 1.6559287183002056, "grad_norm": 0.7587099965919871, "learning_rate": 6.759811247054918e-05, "loss": 1.0304, "step": 302 }, { "epoch": 1.6614119259766964, "grad_norm": 1.6220429069155198, "learning_rate": 6.748684472147255e-05, "loss": 1.0042, "step": 303 }, { "epoch": 1.666895133653187, "grad_norm": 0.9060198165841729, "learning_rate": 6.737517253035027e-05, "loss": 1.0264, "step": 304 }, { "epoch": 1.672378341329678, "grad_norm": 0.7810840322821543, "learning_rate": 6.726309754032942e-05, "loss": 1.0194, "step": 305 }, { "epoch": 1.6778615490061686, "grad_norm": 0.8389807329289671, "learning_rate": 6.715062140048392e-05, "loss": 1.0336, "step": 306 }, { "epoch": 1.6833447566826594, "grad_norm": 0.9811139261812258, "learning_rate": 6.703774576579018e-05, "loss": 0.9961, "step": 307 }, { "epoch": 1.6888279643591502, "grad_norm": 1.693203875036964, "learning_rate": 6.69244722971028e-05, "loss": 1.0054, "step": 308 }, { "epoch": 1.6943111720356407, "grad_norm": 0.6313152362555375, "learning_rate": 6.681080266113017e-05, "loss": 1.0138, "step": 309 }, { "epoch": 1.6997943797121315, "grad_norm": 1.2050041258375503, "learning_rate": 6.669673853040979e-05, "loss": 1.0181, "step": 310 }, { "epoch": 1.7052775873886223, "grad_norm": 1.0692574241468813, "learning_rate": 6.658228158328384e-05, "loss": 1.0196, "step": 311 }, { "epoch": 1.7107607950651131, "grad_norm": 1.3523681454597505, "learning_rate": 6.646743350387438e-05, "loss": 1.0405, "step": 312 }, { "epoch": 1.716244002741604, "grad_norm": 0.9778959208393927, "learning_rate": 6.635219598205863e-05, "loss": 1.0308, "step": 313 }, { "epoch": 1.7217272104180945, "grad_norm": 0.9710639823435132, "learning_rate": 6.623657071344407e-05, "loss": 0.9845, "step": 314 }, { "epoch": 1.7272104180945853, "grad_norm": 1.1795695782951767, "learning_rate": 6.61205593993434e-05, "loss": 1.0354, "step": 315 }, { "epoch": 1.732693625771076, "grad_norm": 1.256640921349798, "learning_rate": 6.600416374674978e-05, "loss": 1.0074, "step": 316 }, { "epoch": 1.7381768334475667, "grad_norm": 0.989814843263136, "learning_rate": 6.588738546831136e-05, "loss": 1.025, "step": 317 }, { "epoch": 1.7436600411240577, "grad_norm": 0.9259711398370861, "learning_rate": 6.577022628230638e-05, "loss": 1.0132, "step": 318 }, { "epoch": 1.7491432488005483, "grad_norm": 0.7510586657590674, "learning_rate": 6.565268791261769e-05, "loss": 1.0092, "step": 319 }, { "epoch": 1.754626456477039, "grad_norm": 0.6249763060539171, "learning_rate": 6.553477208870748e-05, "loss": 1.0198, "step": 320 }, { "epoch": 1.7601096641535299, "grad_norm": 0.6961806925064483, "learning_rate": 6.541648054559182e-05, "loss": 1.0265, "step": 321 }, { "epoch": 1.7655928718300204, "grad_norm": 0.8447275384338139, "learning_rate": 6.529781502381509e-05, "loss": 1.014, "step": 322 }, { "epoch": 1.7710760795065115, "grad_norm": 1.052987904165586, "learning_rate": 6.517877726942445e-05, "loss": 0.9936, "step": 323 }, { "epoch": 1.776559287183002, "grad_norm": 1.2555546445227173, "learning_rate": 6.505936903394406e-05, "loss": 0.9993, "step": 324 }, { "epoch": 1.7820424948594928, "grad_norm": 0.7871410118515071, "learning_rate": 6.493959207434934e-05, "loss": 1.0246, "step": 325 }, { "epoch": 1.7875257025359836, "grad_norm": 0.8254368841706734, "learning_rate": 6.481944815304117e-05, "loss": 0.9871, "step": 326 }, { "epoch": 1.7930089102124742, "grad_norm": 2.389893141482841, "learning_rate": 6.469893903781987e-05, "loss": 1.0288, "step": 327 }, { "epoch": 1.798492117888965, "grad_norm": 0.8529982927275416, "learning_rate": 6.457806650185925e-05, "loss": 1.0209, "step": 328 }, { "epoch": 1.8039753255654558, "grad_norm": 1.9419731761625634, "learning_rate": 6.44568323236805e-05, "loss": 0.9977, "step": 329 }, { "epoch": 1.8094585332419464, "grad_norm": 0.8112015130720743, "learning_rate": 6.433523828712599e-05, "loss": 1.0182, "step": 330 }, { "epoch": 1.8149417409184374, "grad_norm": 1.8647552723000032, "learning_rate": 6.421328618133312e-05, "loss": 1.0059, "step": 331 }, { "epoch": 1.820424948594928, "grad_norm": 1.1202805462423913, "learning_rate": 6.409097780070789e-05, "loss": 1.0315, "step": 332 }, { "epoch": 1.8259081562714188, "grad_norm": 1.4717552883952516, "learning_rate": 6.396831494489852e-05, "loss": 1.0178, "step": 333 }, { "epoch": 1.8313913639479096, "grad_norm": 1.247198837427445, "learning_rate": 6.384529941876902e-05, "loss": 1.0244, "step": 334 }, { "epoch": 1.8368745716244002, "grad_norm": 1.165713928714055, "learning_rate": 6.372193303237258e-05, "loss": 1.0234, "step": 335 }, { "epoch": 1.8423577793008912, "grad_norm": 0.819719891951629, "learning_rate": 6.359821760092493e-05, "loss": 1.0252, "step": 336 }, { "epoch": 1.8478409869773817, "grad_norm": 0.9694295545160206, "learning_rate": 6.347415494477771e-05, "loss": 1.0005, "step": 337 }, { "epoch": 1.8533241946538725, "grad_norm": 0.833369832433467, "learning_rate": 6.334974688939161e-05, "loss": 1.0005, "step": 338 }, { "epoch": 1.8588074023303633, "grad_norm": 1.051580411769736, "learning_rate": 6.322499526530951e-05, "loss": 0.994, "step": 339 }, { "epoch": 1.864290610006854, "grad_norm": 1.3409908447267063, "learning_rate": 6.30999019081296e-05, "loss": 1.0051, "step": 340 }, { "epoch": 1.8697738176833447, "grad_norm": 0.7200296682178966, "learning_rate": 6.297446865847833e-05, "loss": 0.9916, "step": 341 }, { "epoch": 1.8752570253598355, "grad_norm": 0.5703748532613969, "learning_rate": 6.284869736198332e-05, "loss": 1.0117, "step": 342 }, { "epoch": 1.880740233036326, "grad_norm": 0.656233302341983, "learning_rate": 6.272258986924624e-05, "loss": 0.9995, "step": 343 }, { "epoch": 1.8862234407128171, "grad_norm": 0.9959957194372924, "learning_rate": 6.259614803581553e-05, "loss": 0.9863, "step": 344 }, { "epoch": 1.8917066483893077, "grad_norm": 1.2637430599411104, "learning_rate": 6.246937372215916e-05, "loss": 1.0044, "step": 345 }, { "epoch": 1.8971898560657985, "grad_norm": 0.6655587539599807, "learning_rate": 6.23422687936372e-05, "loss": 1.004, "step": 346 }, { "epoch": 1.9026730637422893, "grad_norm": 0.6292486765776564, "learning_rate": 6.22148351204744e-05, "loss": 1.0071, "step": 347 }, { "epoch": 1.9081562714187799, "grad_norm": 0.7336634351921288, "learning_rate": 6.208707457773267e-05, "loss": 0.9866, "step": 348 }, { "epoch": 1.9136394790952709, "grad_norm": 0.6842175616230739, "learning_rate": 6.195898904528346e-05, "loss": 0.9911, "step": 349 }, { "epoch": 1.9191226867717615, "grad_norm": 0.6937670660831978, "learning_rate": 6.183058040778018e-05, "loss": 0.9927, "step": 350 }, { "epoch": 1.9246058944482523, "grad_norm": 0.6943683013295557, "learning_rate": 6.170185055463039e-05, "loss": 0.9883, "step": 351 }, { "epoch": 1.930089102124743, "grad_norm": 0.6759661412127859, "learning_rate": 6.157280137996797e-05, "loss": 0.9792, "step": 352 }, { "epoch": 1.9355723098012336, "grad_norm": 0.7487530675075756, "learning_rate": 6.14434347826254e-05, "loss": 0.9917, "step": 353 }, { "epoch": 1.9410555174777244, "grad_norm": 0.8480627285748816, "learning_rate": 6.131375266610564e-05, "loss": 1.0116, "step": 354 }, { "epoch": 1.9465387251542152, "grad_norm": 1.026881712032747, "learning_rate": 6.118375693855426e-05, "loss": 0.997, "step": 355 }, { "epoch": 1.9520219328307058, "grad_norm": 1.1819824734836144, "learning_rate": 6.10534495127313e-05, "loss": 0.9866, "step": 356 }, { "epoch": 1.9575051405071968, "grad_norm": 0.7488114847350166, "learning_rate": 6.092283230598311e-05, "loss": 0.9992, "step": 357 }, { "epoch": 1.9629883481836874, "grad_norm": 0.5942125372742829, "learning_rate": 6.079190724021418e-05, "loss": 1.0019, "step": 358 }, { "epoch": 1.9684715558601782, "grad_norm": 0.6416782063906716, "learning_rate": 6.066067624185886e-05, "loss": 0.9851, "step": 359 }, { "epoch": 1.973954763536669, "grad_norm": 0.6541261024351206, "learning_rate": 6.0529141241852974e-05, "loss": 0.9967, "step": 360 }, { "epoch": 1.9794379712131596, "grad_norm": 0.7497124499642812, "learning_rate": 6.0397304175605444e-05, "loss": 0.9809, "step": 361 }, { "epoch": 1.9849211788896506, "grad_norm": 0.868567720778104, "learning_rate": 6.026516698296979e-05, "loss": 1.0105, "step": 362 }, { "epoch": 1.9904043865661412, "grad_norm": 0.9302778924884637, "learning_rate": 6.0132731608215626e-05, "loss": 0.9922, "step": 363 }, { "epoch": 1.995887594242632, "grad_norm": 0.9976983084275789, "learning_rate": 6.000000000000001e-05, "loss": 1.0146, "step": 364 }, { "epoch": 2.0013708019191228, "grad_norm": 1.7195064119299488, "learning_rate": 5.9866974111338764e-05, "loss": 1.5667, "step": 365 }, { "epoch": 2.0068540095956133, "grad_norm": 0.8460708264395937, "learning_rate": 5.973365589957777e-05, "loss": 0.9585, "step": 366 }, { "epoch": 2.0123372172721044, "grad_norm": 0.6860754084451314, "learning_rate": 5.96000473263642e-05, "loss": 0.9479, "step": 367 }, { "epoch": 2.017820424948595, "grad_norm": 0.569858067536389, "learning_rate": 5.946615035761756e-05, "loss": 0.9485, "step": 368 }, { "epoch": 2.0233036326250855, "grad_norm": 0.4334842237308715, "learning_rate": 5.9331966963500825e-05, "loss": 0.9368, "step": 369 }, { "epoch": 2.0287868403015765, "grad_norm": 0.5181677147256706, "learning_rate": 5.919749911839146e-05, "loss": 0.9804, "step": 370 }, { "epoch": 2.034270047978067, "grad_norm": 0.6881500556087109, "learning_rate": 5.9062748800852315e-05, "loss": 0.9499, "step": 371 }, { "epoch": 2.039753255654558, "grad_norm": 0.7541551892158753, "learning_rate": 5.892771799360258e-05, "loss": 0.9832, "step": 372 }, { "epoch": 2.0452364633310487, "grad_norm": 0.7791303072747863, "learning_rate": 5.879240868348857e-05, "loss": 0.9449, "step": 373 }, { "epoch": 2.0507196710075393, "grad_norm": 0.7144846911633607, "learning_rate": 5.865682286145446e-05, "loss": 0.9534, "step": 374 }, { "epoch": 2.0562028786840303, "grad_norm": 0.677470076470832, "learning_rate": 5.852096252251308e-05, "loss": 0.9579, "step": 375 }, { "epoch": 2.061686086360521, "grad_norm": 0.7427233443080802, "learning_rate": 5.8384829665716475e-05, "loss": 0.9233, "step": 376 }, { "epoch": 2.0671692940370114, "grad_norm": 0.7777873567557757, "learning_rate": 5.824842629412653e-05, "loss": 0.9706, "step": 377 }, { "epoch": 2.0726525017135025, "grad_norm": 0.7135923513575179, "learning_rate": 5.8111754414785504e-05, "loss": 0.9232, "step": 378 }, { "epoch": 2.078135709389993, "grad_norm": 0.7125391347465736, "learning_rate": 5.797481603868646e-05, "loss": 0.9525, "step": 379 }, { "epoch": 2.083618917066484, "grad_norm": 0.7238619031777882, "learning_rate": 5.783761318074373e-05, "loss": 0.9525, "step": 380 }, { "epoch": 2.0891021247429746, "grad_norm": 0.7128949358918412, "learning_rate": 5.770014785976322e-05, "loss": 0.9511, "step": 381 }, { "epoch": 2.094585332419465, "grad_norm": 0.7209215696044494, "learning_rate": 5.756242209841272e-05, "loss": 0.9636, "step": 382 }, { "epoch": 2.1000685400959562, "grad_norm": 0.6327662060979778, "learning_rate": 5.742443792319216e-05, "loss": 0.968, "step": 383 }, { "epoch": 2.105551747772447, "grad_norm": 0.4985051187489137, "learning_rate": 5.728619736440375e-05, "loss": 0.953, "step": 384 }, { "epoch": 2.111034955448938, "grad_norm": 0.3861018767101939, "learning_rate": 5.714770245612217e-05, "loss": 0.967, "step": 385 }, { "epoch": 2.1165181631254284, "grad_norm": 0.32959266195178644, "learning_rate": 5.700895523616459e-05, "loss": 0.9281, "step": 386 }, { "epoch": 2.122001370801919, "grad_norm": 0.34591568155580654, "learning_rate": 5.6869957746060675e-05, "loss": 0.9521, "step": 387 }, { "epoch": 2.12748457847841, "grad_norm": 0.4625069922403467, "learning_rate": 5.673071203102261e-05, "loss": 0.9628, "step": 388 }, { "epoch": 2.1329677861549006, "grad_norm": 0.4988336639530923, "learning_rate": 5.6591220139914945e-05, "loss": 0.9568, "step": 389 }, { "epoch": 2.138450993831391, "grad_norm": 0.49194383493690574, "learning_rate": 5.645148412522447e-05, "loss": 0.9524, "step": 390 }, { "epoch": 2.143934201507882, "grad_norm": 0.43352665389513567, "learning_rate": 5.6311506043030006e-05, "loss": 0.9469, "step": 391 }, { "epoch": 2.1494174091843727, "grad_norm": 0.3306646386411896, "learning_rate": 5.61712879529722e-05, "loss": 0.9417, "step": 392 }, { "epoch": 2.1549006168608638, "grad_norm": 0.29441855117758814, "learning_rate": 5.6030831918223136e-05, "loss": 0.9579, "step": 393 }, { "epoch": 2.1603838245373543, "grad_norm": 0.3597929142259359, "learning_rate": 5.5890140005456056e-05, "loss": 0.9299, "step": 394 }, { "epoch": 2.165867032213845, "grad_norm": 0.4844488175844035, "learning_rate": 5.574921428481487e-05, "loss": 0.9401, "step": 395 }, { "epoch": 2.171350239890336, "grad_norm": 0.6802142906239217, "learning_rate": 5.5608056829883796e-05, "loss": 0.9357, "step": 396 }, { "epoch": 2.1768334475668265, "grad_norm": 1.2106002196981427, "learning_rate": 5.546666971765675e-05, "loss": 0.9651, "step": 397 }, { "epoch": 2.1823166552433175, "grad_norm": 0.6157717006727198, "learning_rate": 5.532505502850688e-05, "loss": 0.9528, "step": 398 }, { "epoch": 2.187799862919808, "grad_norm": 0.641804078912436, "learning_rate": 5.5183214846155864e-05, "loss": 0.9718, "step": 399 }, { "epoch": 2.1932830705962987, "grad_norm": 0.6070401599275227, "learning_rate": 5.504115125764329e-05, "loss": 0.9572, "step": 400 }, { "epoch": 2.1987662782727897, "grad_norm": 0.7100305243791368, "learning_rate": 5.489886635329598e-05, "loss": 0.9417, "step": 401 }, { "epoch": 2.2042494859492803, "grad_norm": 0.9164307400605365, "learning_rate": 5.4756362226697193e-05, "loss": 0.9449, "step": 402 }, { "epoch": 2.209732693625771, "grad_norm": 1.3336341868468473, "learning_rate": 5.461364097465581e-05, "loss": 0.9758, "step": 403 }, { "epoch": 2.215215901302262, "grad_norm": 0.5707410680573715, "learning_rate": 5.447070469717552e-05, "loss": 0.9509, "step": 404 }, { "epoch": 2.2206991089787524, "grad_norm": 0.5349427819988657, "learning_rate": 5.4327555497423874e-05, "loss": 0.977, "step": 405 }, { "epoch": 2.2261823166552435, "grad_norm": 0.6665137379275449, "learning_rate": 5.4184195481701425e-05, "loss": 0.9718, "step": 406 }, { "epoch": 2.231665524331734, "grad_norm": 0.6788016902611529, "learning_rate": 5.4040626759410625e-05, "loss": 0.9559, "step": 407 }, { "epoch": 2.2371487320082246, "grad_norm": 0.7062646461019303, "learning_rate": 5.3896851443024837e-05, "loss": 0.9663, "step": 408 }, { "epoch": 2.2426319396847156, "grad_norm": 0.6871839246478715, "learning_rate": 5.375287164805727e-05, "loss": 0.9814, "step": 409 }, { "epoch": 2.248115147361206, "grad_norm": 0.5968021041080723, "learning_rate": 5.360868949302986e-05, "loss": 0.937, "step": 410 }, { "epoch": 2.2535983550376972, "grad_norm": 0.5805892891972629, "learning_rate": 5.3464307099442035e-05, "loss": 0.9479, "step": 411 }, { "epoch": 2.259081562714188, "grad_norm": 0.4860199686870337, "learning_rate": 5.3319726591739536e-05, "loss": 0.9541, "step": 412 }, { "epoch": 2.2645647703906784, "grad_norm": 0.5697676557756951, "learning_rate": 5.317495009728319e-05, "loss": 0.9471, "step": 413 }, { "epoch": 2.2700479780671694, "grad_norm": 0.3474403252554382, "learning_rate": 5.302997974631757e-05, "loss": 0.9535, "step": 414 }, { "epoch": 2.27553118574366, "grad_norm": 0.3892967631879122, "learning_rate": 5.288481767193963e-05, "loss": 0.9573, "step": 415 }, { "epoch": 2.2810143934201506, "grad_norm": 0.43748961608691317, "learning_rate": 5.2739466010067385e-05, "loss": 0.9377, "step": 416 }, { "epoch": 2.2864976010966416, "grad_norm": 0.410672984619657, "learning_rate": 5.259392689940841e-05, "loss": 0.9313, "step": 417 }, { "epoch": 2.291980808773132, "grad_norm": 0.44449306463157623, "learning_rate": 5.244820248142844e-05, "loss": 0.9373, "step": 418 }, { "epoch": 2.297464016449623, "grad_norm": 0.457420584226533, "learning_rate": 5.2302294900319796e-05, "loss": 0.9627, "step": 419 }, { "epoch": 2.3029472241261137, "grad_norm": 0.49606181089497686, "learning_rate": 5.215620630296988e-05, "loss": 0.9491, "step": 420 }, { "epoch": 2.3084304318026048, "grad_norm": 0.5557714109621562, "learning_rate": 5.200993883892956e-05, "loss": 0.9538, "step": 421 }, { "epoch": 2.3139136394790953, "grad_norm": 0.5848899773483123, "learning_rate": 5.1863494660381586e-05, "loss": 0.9589, "step": 422 }, { "epoch": 2.319396847155586, "grad_norm": 0.5756814498770187, "learning_rate": 5.1716875922108836e-05, "loss": 0.9437, "step": 423 }, { "epoch": 2.324880054832077, "grad_norm": 0.5975395685498862, "learning_rate": 5.1570084781462716e-05, "loss": 0.9752, "step": 424 }, { "epoch": 2.3303632625085675, "grad_norm": 0.6741079069902028, "learning_rate": 5.142312339833131e-05, "loss": 0.9695, "step": 425 }, { "epoch": 2.335846470185058, "grad_norm": 0.47322117724121776, "learning_rate": 5.1275993935107714e-05, "loss": 0.9376, "step": 426 }, { "epoch": 2.341329677861549, "grad_norm": 0.3713624782507764, "learning_rate": 5.112869855665811e-05, "loss": 0.9587, "step": 427 }, { "epoch": 2.3468128855380397, "grad_norm": 0.3297040646564409, "learning_rate": 5.098123943028999e-05, "loss": 0.95, "step": 428 }, { "epoch": 2.3522960932145303, "grad_norm": 0.3283051265027244, "learning_rate": 5.0833618725720214e-05, "loss": 0.9585, "step": 429 }, { "epoch": 2.3577793008910213, "grad_norm": 0.33355324970835326, "learning_rate": 5.0685838615043124e-05, "loss": 0.9373, "step": 430 }, { "epoch": 2.363262508567512, "grad_norm": 0.32077316347401524, "learning_rate": 5.053790127269855e-05, "loss": 0.9281, "step": 431 }, { "epoch": 2.368745716244003, "grad_norm": 0.29194427533135364, "learning_rate": 5.038980887543987e-05, "loss": 0.9565, "step": 432 }, { "epoch": 2.3742289239204935, "grad_norm": 0.2613277032435807, "learning_rate": 5.024156360230189e-05, "loss": 0.9654, "step": 433 }, { "epoch": 2.3797121315969845, "grad_norm": 0.30838651942670436, "learning_rate": 5.0093167634568874e-05, "loss": 0.9555, "step": 434 }, { "epoch": 2.385195339273475, "grad_norm": 0.3084305013600498, "learning_rate": 4.9944623155742395e-05, "loss": 0.9301, "step": 435 }, { "epoch": 2.3906785469499656, "grad_norm": 0.3387529105331699, "learning_rate": 4.979593235150924e-05, "loss": 0.9428, "step": 436 }, { "epoch": 2.3961617546264566, "grad_norm": 0.3995434200366217, "learning_rate": 4.9647097409709186e-05, "loss": 0.9405, "step": 437 }, { "epoch": 2.401644962302947, "grad_norm": 0.41290791299928126, "learning_rate": 4.94981205203029e-05, "loss": 0.9483, "step": 438 }, { "epoch": 2.407128169979438, "grad_norm": 0.4272879341834311, "learning_rate": 4.934900387533965e-05, "loss": 0.9434, "step": 439 }, { "epoch": 2.412611377655929, "grad_norm": 0.46378214773916265, "learning_rate": 4.9199749668925076e-05, "loss": 0.9551, "step": 440 }, { "epoch": 2.4180945853324194, "grad_norm": 0.46144056548958406, "learning_rate": 4.9050360097188904e-05, "loss": 0.952, "step": 441 }, { "epoch": 2.42357779300891, "grad_norm": 0.4496399315688108, "learning_rate": 4.890083735825258e-05, "loss": 0.9414, "step": 442 }, { "epoch": 2.429061000685401, "grad_norm": 0.5404123880896423, "learning_rate": 4.875118365219706e-05, "loss": 0.976, "step": 443 }, { "epoch": 2.4345442083618916, "grad_norm": 0.6694234081039024, "learning_rate": 4.86014011810303e-05, "loss": 0.938, "step": 444 }, { "epoch": 2.4400274160383826, "grad_norm": 0.7976721444374203, "learning_rate": 4.845149214865491e-05, "loss": 0.9428, "step": 445 }, { "epoch": 2.445510623714873, "grad_norm": 0.8514351115579054, "learning_rate": 4.830145876083575e-05, "loss": 0.9716, "step": 446 }, { "epoch": 2.450993831391364, "grad_norm": 0.8419233920727014, "learning_rate": 4.81513032251674e-05, "loss": 0.9375, "step": 447 }, { "epoch": 2.4564770390678548, "grad_norm": 0.7939704221285054, "learning_rate": 4.8001027751041784e-05, "loss": 0.9539, "step": 448 }, { "epoch": 2.4619602467443453, "grad_norm": 0.6506959170383838, "learning_rate": 4.785063454961557e-05, "loss": 0.974, "step": 449 }, { "epoch": 2.4674434544208363, "grad_norm": 0.4039955860724156, "learning_rate": 4.7700125833777664e-05, "loss": 0.9409, "step": 450 }, { "epoch": 2.472926662097327, "grad_norm": 0.33076049398826807, "learning_rate": 4.754950381811667e-05, "loss": 0.9561, "step": 451 }, { "epoch": 2.4784098697738175, "grad_norm": 0.4841447525816379, "learning_rate": 4.7398770718888296e-05, "loss": 0.9557, "step": 452 }, { "epoch": 2.4838930774503085, "grad_norm": 0.6029861941124044, "learning_rate": 4.724792875398271e-05, "loss": 0.9254, "step": 453 }, { "epoch": 2.489376285126799, "grad_norm": 0.6018267599600162, "learning_rate": 4.7096980142891936e-05, "loss": 0.9336, "step": 454 }, { "epoch": 2.49485949280329, "grad_norm": 0.4547177146031252, "learning_rate": 4.694592710667723e-05, "loss": 0.941, "step": 455 }, { "epoch": 2.5003427004797807, "grad_norm": 0.2560704544963682, "learning_rate": 4.6794771867936286e-05, "loss": 0.9413, "step": 456 }, { "epoch": 2.5058259081562713, "grad_norm": 0.30599482861554783, "learning_rate": 4.66435166507707e-05, "loss": 0.9605, "step": 457 }, { "epoch": 2.5113091158327623, "grad_norm": 0.42429824249004056, "learning_rate": 4.6492163680753096e-05, "loss": 0.9403, "step": 458 }, { "epoch": 2.516792323509253, "grad_norm": 0.476594932388952, "learning_rate": 4.634071518489443e-05, "loss": 0.9354, "step": 459 }, { "epoch": 2.522275531185744, "grad_norm": 0.46789821860796127, "learning_rate": 4.618917339161125e-05, "loss": 0.9495, "step": 460 }, { "epoch": 2.5277587388622345, "grad_norm": 0.3784567986912495, "learning_rate": 4.6037540530692905e-05, "loss": 0.958, "step": 461 }, { "epoch": 2.533241946538725, "grad_norm": 0.3589933646618808, "learning_rate": 4.588581883326865e-05, "loss": 0.9449, "step": 462 }, { "epoch": 2.538725154215216, "grad_norm": 0.4378215691524641, "learning_rate": 4.573401053177494e-05, "loss": 0.949, "step": 463 }, { "epoch": 2.5442083618917066, "grad_norm": 0.4673331119883799, "learning_rate": 4.558211785992251e-05, "loss": 0.9555, "step": 464 }, { "epoch": 2.549691569568197, "grad_norm": 0.4792027471063911, "learning_rate": 4.543014305266352e-05, "loss": 0.981, "step": 465 }, { "epoch": 2.5551747772446882, "grad_norm": 0.4774897666673642, "learning_rate": 4.5278088346158665e-05, "loss": 0.9456, "step": 466 }, { "epoch": 2.560657984921179, "grad_norm": 0.4425042364504943, "learning_rate": 4.512595597774427e-05, "loss": 0.9429, "step": 467 }, { "epoch": 2.5661411925976694, "grad_norm": 0.39911527326216606, "learning_rate": 4.4973748185899416e-05, "loss": 0.9413, "step": 468 }, { "epoch": 2.5716244002741604, "grad_norm": 0.29210171265761314, "learning_rate": 4.4821467210212924e-05, "loss": 0.9297, "step": 469 }, { "epoch": 2.577107607950651, "grad_norm": 0.25565384930514384, "learning_rate": 4.4669115291350484e-05, "loss": 0.9454, "step": 470 }, { "epoch": 2.582590815627142, "grad_norm": 0.2581461861152204, "learning_rate": 4.451669467102162e-05, "loss": 0.948, "step": 471 }, { "epoch": 2.5880740233036326, "grad_norm": 0.29491880993583836, "learning_rate": 4.436420759194671e-05, "loss": 0.9381, "step": 472 }, { "epoch": 2.5935572309801236, "grad_norm": 0.31658803555973, "learning_rate": 4.4211656297824064e-05, "loss": 0.9618, "step": 473 }, { "epoch": 2.599040438656614, "grad_norm": 0.3245542687115596, "learning_rate": 4.4059043033296815e-05, "loss": 0.9488, "step": 474 }, { "epoch": 2.6045236463331047, "grad_norm": 0.2905774382505926, "learning_rate": 4.390637004391993e-05, "loss": 0.9323, "step": 475 }, { "epoch": 2.6100068540095958, "grad_norm": 0.2428082128722324, "learning_rate": 4.375363957612717e-05, "loss": 0.9549, "step": 476 }, { "epoch": 2.6154900616860863, "grad_norm": 0.24550826505798423, "learning_rate": 4.360085387719806e-05, "loss": 0.9493, "step": 477 }, { "epoch": 2.620973269362577, "grad_norm": 0.30222793473570203, "learning_rate": 4.344801519522478e-05, "loss": 0.9317, "step": 478 }, { "epoch": 2.626456477039068, "grad_norm": 0.3497976704406349, "learning_rate": 4.32951257790791e-05, "loss": 0.9377, "step": 479 }, { "epoch": 2.6319396847155585, "grad_norm": 1.2304346438090321, "learning_rate": 4.314218787837925e-05, "loss": 0.9442, "step": 480 }, { "epoch": 2.637422892392049, "grad_norm": 0.2466343919140661, "learning_rate": 4.298920374345698e-05, "loss": 0.944, "step": 481 }, { "epoch": 2.64290610006854, "grad_norm": 1.8185754976050628, "learning_rate": 4.283617562532421e-05, "loss": 0.9825, "step": 482 }, { "epoch": 2.648389307745031, "grad_norm": 0.5797415336422694, "learning_rate": 4.2683105775640096e-05, "loss": 0.9565, "step": 483 }, { "epoch": 2.6538725154215217, "grad_norm": 0.6585844779419097, "learning_rate": 4.2529996446677814e-05, "loss": 0.9349, "step": 484 }, { "epoch": 2.6593557230980123, "grad_norm": 0.6364339320143122, "learning_rate": 4.237684989129146e-05, "loss": 0.9407, "step": 485 }, { "epoch": 2.6648389307745033, "grad_norm": 0.7003736753475072, "learning_rate": 4.2223668362882846e-05, "loss": 0.952, "step": 486 }, { "epoch": 2.670322138450994, "grad_norm": 0.6868570924103546, "learning_rate": 4.2070454115368385e-05, "loss": 0.9465, "step": 487 }, { "epoch": 2.6758053461274844, "grad_norm": 0.8377123241663084, "learning_rate": 4.191720940314593e-05, "loss": 0.9365, "step": 488 }, { "epoch": 2.6812885538039755, "grad_norm": 1.6613962498279427, "learning_rate": 4.176393648106161e-05, "loss": 0.9592, "step": 489 }, { "epoch": 2.686771761480466, "grad_norm": 0.9696773125080363, "learning_rate": 4.1610637604376614e-05, "loss": 0.9526, "step": 490 }, { "epoch": 2.6922549691569566, "grad_norm": 1.0222168588069447, "learning_rate": 4.1457315028734015e-05, "loss": 0.9736, "step": 491 }, { "epoch": 2.6977381768334476, "grad_norm": 0.9553354559606164, "learning_rate": 4.13039710101256e-05, "loss": 0.9668, "step": 492 }, { "epoch": 2.703221384509938, "grad_norm": 0.8745999789478393, "learning_rate": 4.11506078048587e-05, "loss": 0.9716, "step": 493 }, { "epoch": 2.7087045921864292, "grad_norm": 0.7411939891535938, "learning_rate": 4.0997227669522924e-05, "loss": 0.9602, "step": 494 }, { "epoch": 2.71418779986292, "grad_norm": 0.6053293148082218, "learning_rate": 4.0843832860956994e-05, "loss": 0.9321, "step": 495 }, { "epoch": 2.719671007539411, "grad_norm": 0.6015174283144377, "learning_rate": 4.069042563621555e-05, "loss": 0.9556, "step": 496 }, { "epoch": 2.7251542152159014, "grad_norm": 0.5341988781830062, "learning_rate": 4.0537008252535904e-05, "loss": 0.9502, "step": 497 }, { "epoch": 2.730637422892392, "grad_norm": 0.3986261784888644, "learning_rate": 4.0383582967304865e-05, "loss": 0.9498, "step": 498 }, { "epoch": 2.736120630568883, "grad_norm": 0.4491497759193806, "learning_rate": 4.023015203802551e-05, "loss": 0.9578, "step": 499 }, { "epoch": 2.7416038382453736, "grad_norm": 0.42597854684305547, "learning_rate": 4.0076717722283936e-05, "loss": 0.9258, "step": 500 }, { "epoch": 2.747087045921864, "grad_norm": 0.3674427892141133, "learning_rate": 3.992328227771608e-05, "loss": 0.956, "step": 501 }, { "epoch": 2.752570253598355, "grad_norm": 0.3441073888754271, "learning_rate": 3.976984796197451e-05, "loss": 0.9347, "step": 502 }, { "epoch": 2.7580534612748457, "grad_norm": 0.5779083792402964, "learning_rate": 3.961641703269514e-05, "loss": 0.9591, "step": 503 }, { "epoch": 2.7635366689513363, "grad_norm": 0.45842637741331516, "learning_rate": 3.946299174746411e-05, "loss": 0.9497, "step": 504 }, { "epoch": 2.7690198766278273, "grad_norm": 0.42003328606156265, "learning_rate": 3.9309574363784465e-05, "loss": 0.9365, "step": 505 }, { "epoch": 2.774503084304318, "grad_norm": 0.3971675276536041, "learning_rate": 3.915616713904302e-05, "loss": 0.9532, "step": 506 }, { "epoch": 2.779986291980809, "grad_norm": 0.3738039726781753, "learning_rate": 3.9002772330477096e-05, "loss": 0.9505, "step": 507 }, { "epoch": 2.7854694996572995, "grad_norm": 0.3490205849659376, "learning_rate": 3.884939219514132e-05, "loss": 0.9383, "step": 508 }, { "epoch": 2.7909527073337905, "grad_norm": 0.6333313596692628, "learning_rate": 3.869602898987441e-05, "loss": 0.9528, "step": 509 }, { "epoch": 2.796435915010281, "grad_norm": 0.3242106302112962, "learning_rate": 3.854268497126601e-05, "loss": 0.9554, "step": 510 }, { "epoch": 2.8019191226867717, "grad_norm": 0.2849035104268222, "learning_rate": 3.8389362395623406e-05, "loss": 0.9477, "step": 511 }, { "epoch": 2.8074023303632627, "grad_norm": 0.23339126446276168, "learning_rate": 3.8236063518938405e-05, "loss": 0.9401, "step": 512 }, { "epoch": 2.8128855380397533, "grad_norm": 0.2658559032559636, "learning_rate": 3.8082790596854075e-05, "loss": 0.9322, "step": 513 }, { "epoch": 2.818368745716244, "grad_norm": 0.28632019300770967, "learning_rate": 3.792954588463162e-05, "loss": 0.9647, "step": 514 }, { "epoch": 2.823851953392735, "grad_norm": 0.2585763044971737, "learning_rate": 3.777633163711716e-05, "loss": 0.9505, "step": 515 }, { "epoch": 2.8293351610692254, "grad_norm": 0.2742118891600389, "learning_rate": 3.7623150108708546e-05, "loss": 0.9446, "step": 516 }, { "epoch": 2.834818368745716, "grad_norm": 0.2344118819440086, "learning_rate": 3.7470003553322186e-05, "loss": 0.9309, "step": 517 }, { "epoch": 2.840301576422207, "grad_norm": 0.5837521053187771, "learning_rate": 3.7316894224359904e-05, "loss": 0.9394, "step": 518 }, { "epoch": 2.8457847840986976, "grad_norm": 0.31498022974880685, "learning_rate": 3.71638243746758e-05, "loss": 0.9321, "step": 519 }, { "epoch": 2.8512679917751886, "grad_norm": 0.5111390953286222, "learning_rate": 3.7010796256543034e-05, "loss": 0.9565, "step": 520 }, { "epoch": 2.856751199451679, "grad_norm": 0.24452242688766815, "learning_rate": 3.6857812121620756e-05, "loss": 0.938, "step": 521 }, { "epoch": 2.8622344071281702, "grad_norm": 0.263705639317996, "learning_rate": 3.670487422092092e-05, "loss": 0.9434, "step": 522 }, { "epoch": 2.867717614804661, "grad_norm": 0.25074905428863853, "learning_rate": 3.655198480477523e-05, "loss": 0.9514, "step": 523 }, { "epoch": 2.8732008224811514, "grad_norm": 0.23134392999771497, "learning_rate": 3.639914612280194e-05, "loss": 0.9437, "step": 524 }, { "epoch": 2.8786840301576424, "grad_norm": 0.34358764407986175, "learning_rate": 3.6246360423872834e-05, "loss": 0.9518, "step": 525 }, { "epoch": 2.884167237834133, "grad_norm": 0.21629316436514154, "learning_rate": 3.609362995608008e-05, "loss": 0.9572, "step": 526 }, { "epoch": 2.8896504455106236, "grad_norm": 0.22317832298096357, "learning_rate": 3.59409569667032e-05, "loss": 0.9492, "step": 527 }, { "epoch": 2.8951336531871146, "grad_norm": 0.20759859863486543, "learning_rate": 3.578834370217595e-05, "loss": 0.9355, "step": 528 }, { "epoch": 2.900616860863605, "grad_norm": 0.2138741670103374, "learning_rate": 3.5635792408053304e-05, "loss": 0.9526, "step": 529 }, { "epoch": 2.9061000685400957, "grad_norm": 0.2071241548771815, "learning_rate": 3.54833053289784e-05, "loss": 0.9459, "step": 530 }, { "epoch": 2.9115832762165867, "grad_norm": 0.24989924945583794, "learning_rate": 3.533088470864953e-05, "loss": 0.9402, "step": 531 }, { "epoch": 2.9170664838930773, "grad_norm": 0.27509189913976284, "learning_rate": 3.517853278978708e-05, "loss": 0.9554, "step": 532 }, { "epoch": 2.9225496915695683, "grad_norm": 0.23063790000408171, "learning_rate": 3.5026251814100604e-05, "loss": 0.9457, "step": 533 }, { "epoch": 2.928032899246059, "grad_norm": 0.2166555423682174, "learning_rate": 3.487404402225574e-05, "loss": 0.9272, "step": 534 }, { "epoch": 2.93351610692255, "grad_norm": 0.21899933270897742, "learning_rate": 3.4721911653841355e-05, "loss": 0.9507, "step": 535 }, { "epoch": 2.9389993145990405, "grad_norm": 0.19354444875830706, "learning_rate": 3.45698569473365e-05, "loss": 0.9682, "step": 536 }, { "epoch": 2.944482522275531, "grad_norm": 0.17844494801443178, "learning_rate": 3.44178821400775e-05, "loss": 0.9473, "step": 537 }, { "epoch": 2.949965729952022, "grad_norm": 1.3083812330864948, "learning_rate": 3.426598946822507e-05, "loss": 0.9461, "step": 538 }, { "epoch": 2.9554489376285127, "grad_norm": 0.21349382766535052, "learning_rate": 3.4114181166731355e-05, "loss": 0.9238, "step": 539 }, { "epoch": 2.9609321453050033, "grad_norm": 0.2740853283765663, "learning_rate": 3.39624594693071e-05, "loss": 0.9483, "step": 540 }, { "epoch": 2.9664153529814943, "grad_norm": 0.2604547593626456, "learning_rate": 3.381082660838875e-05, "loss": 0.9566, "step": 541 }, { "epoch": 2.971898560657985, "grad_norm": 0.2636302620609088, "learning_rate": 3.365928481510558e-05, "loss": 0.9529, "step": 542 }, { "epoch": 2.9773817683344754, "grad_norm": 0.24591140742789205, "learning_rate": 3.3507836319246924e-05, "loss": 0.9327, "step": 543 }, { "epoch": 2.9828649760109665, "grad_norm": 0.23336910029530064, "learning_rate": 3.33564833492293e-05, "loss": 0.9606, "step": 544 }, { "epoch": 2.988348183687457, "grad_norm": 0.215970712690598, "learning_rate": 3.3205228132063714e-05, "loss": 0.9495, "step": 545 }, { "epoch": 2.993831391363948, "grad_norm": 0.23530537682003938, "learning_rate": 3.305407289332279e-05, "loss": 0.9533, "step": 546 }, { "epoch": 2.9993145990404386, "grad_norm": 0.35138852092483364, "learning_rate": 3.290301985710807e-05, "loss": 1.3456, "step": 547 }, { "epoch": 3.004797806716929, "grad_norm": 0.5481962224464548, "learning_rate": 3.27520712460173e-05, "loss": 1.1027, "step": 548 }, { "epoch": 3.01028101439342, "grad_norm": 0.7662140587544488, "learning_rate": 3.260122928111172e-05, "loss": 0.8979, "step": 549 }, { "epoch": 3.015764222069911, "grad_norm": 0.9454349543994065, "learning_rate": 3.245049618188334e-05, "loss": 0.9039, "step": 550 }, { "epoch": 3.021247429746402, "grad_norm": 0.7559182562175768, "learning_rate": 3.229987416622235e-05, "loss": 0.9046, "step": 551 }, { "epoch": 3.0267306374228924, "grad_norm": 0.6019789035769177, "learning_rate": 3.2149365450384445e-05, "loss": 0.8982, "step": 552 }, { "epoch": 3.032213845099383, "grad_norm": 0.608624397941904, "learning_rate": 3.199897224895823e-05, "loss": 0.9024, "step": 553 }, { "epoch": 3.037697052775874, "grad_norm": 1.3331460907440373, "learning_rate": 3.184869677483261e-05, "loss": 0.9183, "step": 554 }, { "epoch": 3.0431802604523646, "grad_norm": 0.5617254798828216, "learning_rate": 3.169854123916426e-05, "loss": 0.9044, "step": 555 }, { "epoch": 3.0486634681288556, "grad_norm": 0.6117567607126662, "learning_rate": 3.1548507851345094e-05, "loss": 0.9084, "step": 556 }, { "epoch": 3.054146675805346, "grad_norm": 0.5486288335887587, "learning_rate": 3.139859881896971e-05, "loss": 0.9086, "step": 557 }, { "epoch": 3.0596298834818367, "grad_norm": 0.43883951251804787, "learning_rate": 3.124881634780295e-05, "loss": 0.8822, "step": 558 }, { "epoch": 3.0651130911583278, "grad_norm": 0.375243221483072, "learning_rate": 3.109916264174743e-05, "loss": 0.9089, "step": 559 }, { "epoch": 3.0705962988348183, "grad_norm": 0.45192811672936717, "learning_rate": 3.094963990281112e-05, "loss": 0.8979, "step": 560 }, { "epoch": 3.076079506511309, "grad_norm": 0.3691771282212795, "learning_rate": 3.080025033107494e-05, "loss": 0.8973, "step": 561 }, { "epoch": 3.0815627141878, "grad_norm": 0.28371809189511954, "learning_rate": 3.065099612466037e-05, "loss": 0.887, "step": 562 }, { "epoch": 3.0870459218642905, "grad_norm": 0.33397747859909344, "learning_rate": 3.0501879479697112e-05, "loss": 0.9324, "step": 563 }, { "epoch": 3.0925291295407815, "grad_norm": 8.830372813312158, "learning_rate": 3.035290259029083e-05, "loss": 0.9137, "step": 564 }, { "epoch": 3.098012337217272, "grad_norm": 0.6793506372960832, "learning_rate": 3.0204067648490766e-05, "loss": 0.9178, "step": 565 }, { "epoch": 3.1034955448937627, "grad_norm": 0.8864777488999107, "learning_rate": 3.00553768442576e-05, "loss": 0.8926, "step": 566 }, { "epoch": 3.1089787525702537, "grad_norm": 0.6288925135464869, "learning_rate": 2.9906832365431132e-05, "loss": 0.9198, "step": 567 }, { "epoch": 3.1144619602467443, "grad_norm": 0.40885944277256037, "learning_rate": 2.9758436397698118e-05, "loss": 0.9127, "step": 568 }, { "epoch": 3.1199451679232353, "grad_norm": 0.7284002718913494, "learning_rate": 2.961019112456014e-05, "loss": 0.882, "step": 569 }, { "epoch": 3.125428375599726, "grad_norm": 0.5837455202754323, "learning_rate": 2.946209872730145e-05, "loss": 0.9078, "step": 570 }, { "epoch": 3.1309115832762164, "grad_norm": 0.33186964600527347, "learning_rate": 2.931416138495689e-05, "loss": 0.8925, "step": 571 }, { "epoch": 3.1363947909527075, "grad_norm": 0.53882269291948, "learning_rate": 2.9166381274279803e-05, "loss": 0.9042, "step": 572 }, { "epoch": 3.141877998629198, "grad_norm": 0.46331612230401065, "learning_rate": 2.901876056971002e-05, "loss": 0.9217, "step": 573 }, { "epoch": 3.147361206305689, "grad_norm": 0.31771843320448484, "learning_rate": 2.88713014433419e-05, "loss": 0.9018, "step": 574 }, { "epoch": 3.1528444139821796, "grad_norm": 0.49472378335259276, "learning_rate": 2.8724006064892296e-05, "loss": 0.8748, "step": 575 }, { "epoch": 3.15832762165867, "grad_norm": 0.3530540111863458, "learning_rate": 2.85768766016687e-05, "loss": 0.8957, "step": 576 }, { "epoch": 3.1638108293351612, "grad_norm": 0.3023194345015207, "learning_rate": 2.8429915218537297e-05, "loss": 0.9023, "step": 577 }, { "epoch": 3.169294037011652, "grad_norm": 0.39598120224098793, "learning_rate": 2.8283124077891167e-05, "loss": 0.8928, "step": 578 }, { "epoch": 3.1747772446881424, "grad_norm": 0.2762440885300842, "learning_rate": 2.813650533961843e-05, "loss": 0.8986, "step": 579 }, { "epoch": 3.1802604523646334, "grad_norm": 0.2731980966351029, "learning_rate": 2.7990061161070445e-05, "loss": 0.9012, "step": 580 }, { "epoch": 3.185743660041124, "grad_norm": 0.41348684607047675, "learning_rate": 2.7843793697030128e-05, "loss": 0.9069, "step": 581 }, { "epoch": 3.191226867717615, "grad_norm": 0.25162826940534694, "learning_rate": 2.7697705099680217e-05, "loss": 0.9061, "step": 582 }, { "epoch": 3.1967100753941056, "grad_norm": 0.2506514671680193, "learning_rate": 2.7551797518571573e-05, "loss": 0.8851, "step": 583 }, { "epoch": 3.202193283070596, "grad_norm": 0.2659325922311813, "learning_rate": 2.7406073100591605e-05, "loss": 0.8999, "step": 584 }, { "epoch": 3.207676490747087, "grad_norm": 0.22833189604550308, "learning_rate": 2.7260533989932628e-05, "loss": 0.8804, "step": 585 }, { "epoch": 3.2131596984235777, "grad_norm": 0.26002377936612986, "learning_rate": 2.7115182328060385e-05, "loss": 0.8987, "step": 586 }, { "epoch": 3.2186429061000688, "grad_norm": 0.24416881973039506, "learning_rate": 2.697002025368245e-05, "loss": 0.8982, "step": 587 }, { "epoch": 3.2241261137765593, "grad_norm": 0.22174095398609206, "learning_rate": 2.682504990271682e-05, "loss": 0.8962, "step": 588 }, { "epoch": 3.22960932145305, "grad_norm": 0.2715414538357121, "learning_rate": 2.668027340826048e-05, "loss": 0.8896, "step": 589 }, { "epoch": 3.235092529129541, "grad_norm": 0.2765507918752038, "learning_rate": 2.653569290055799e-05, "loss": 0.8938, "step": 590 }, { "epoch": 3.2405757368060315, "grad_norm": 0.20005217231999928, "learning_rate": 2.6391310506970147e-05, "loss": 0.8827, "step": 591 }, { "epoch": 3.246058944482522, "grad_norm": 0.8704396300357945, "learning_rate": 2.6247128351942726e-05, "loss": 0.9075, "step": 592 }, { "epoch": 3.251542152159013, "grad_norm": 0.3586781432686027, "learning_rate": 2.6103148556975173e-05, "loss": 0.8919, "step": 593 }, { "epoch": 3.2570253598355037, "grad_norm": 0.3879500657373189, "learning_rate": 2.5959373240589382e-05, "loss": 0.906, "step": 594 }, { "epoch": 3.2625085675119947, "grad_norm": 0.3206624742458832, "learning_rate": 2.5815804518298575e-05, "loss": 0.9046, "step": 595 }, { "epoch": 3.2679917751884853, "grad_norm": 0.5655322768937269, "learning_rate": 2.5672444502576122e-05, "loss": 0.8792, "step": 596 }, { "epoch": 3.273474982864976, "grad_norm": 0.26423994402378853, "learning_rate": 2.55292953028245e-05, "loss": 0.9029, "step": 597 }, { "epoch": 3.278958190541467, "grad_norm": 0.3568680800314573, "learning_rate": 2.53863590253442e-05, "loss": 0.9097, "step": 598 }, { "epoch": 3.2844413982179574, "grad_norm": 0.2876917661506257, "learning_rate": 2.5243637773302817e-05, "loss": 0.9039, "step": 599 }, { "epoch": 3.2899246058944485, "grad_norm": 0.2543858700320745, "learning_rate": 2.510113364670403e-05, "loss": 0.8974, "step": 600 }, { "epoch": 3.295407813570939, "grad_norm": 0.29660460583308934, "learning_rate": 2.4958848742356724e-05, "loss": 0.9085, "step": 601 }, { "epoch": 3.3008910212474296, "grad_norm": 0.21363483198577174, "learning_rate": 2.481678515384415e-05, "loss": 0.9022, "step": 602 }, { "epoch": 3.3063742289239206, "grad_norm": 0.2406946488569973, "learning_rate": 2.4674944971493123e-05, "loss": 0.9017, "step": 603 }, { "epoch": 3.311857436600411, "grad_norm": 0.25257504290451166, "learning_rate": 2.453333028234325e-05, "loss": 0.913, "step": 604 }, { "epoch": 3.317340644276902, "grad_norm": 0.17923139676357774, "learning_rate": 2.439194317011622e-05, "loss": 0.9097, "step": 605 }, { "epoch": 3.322823851953393, "grad_norm": 0.2282134057524544, "learning_rate": 2.4250785715185138e-05, "loss": 0.8902, "step": 606 }, { "epoch": 3.3283070596298834, "grad_norm": 0.19939767203864542, "learning_rate": 2.410985999454396e-05, "loss": 0.8929, "step": 607 }, { "epoch": 3.3337902673063744, "grad_norm": 0.1962097138596388, "learning_rate": 2.3969168081776867e-05, "loss": 0.8816, "step": 608 }, { "epoch": 3.339273474982865, "grad_norm": 0.2408654729001506, "learning_rate": 2.382871204702781e-05, "loss": 0.8904, "step": 609 }, { "epoch": 3.3447566826593556, "grad_norm": 0.1857502171735821, "learning_rate": 2.3688493956969997e-05, "loss": 0.9253, "step": 610 }, { "epoch": 3.3502398903358466, "grad_norm": 0.1967645774870265, "learning_rate": 2.3548515874775547e-05, "loss": 0.8996, "step": 611 }, { "epoch": 3.355723098012337, "grad_norm": 0.19126170344434232, "learning_rate": 2.340877986008507e-05, "loss": 0.9036, "step": 612 }, { "epoch": 3.361206305688828, "grad_norm": 0.17261773740991576, "learning_rate": 2.3269287968977406e-05, "loss": 0.8974, "step": 613 }, { "epoch": 3.3666895133653187, "grad_norm": 0.1828461955951697, "learning_rate": 2.3130042253939334e-05, "loss": 0.9053, "step": 614 }, { "epoch": 3.3721727210418093, "grad_norm": 0.1787029837346015, "learning_rate": 2.2991044763835438e-05, "loss": 0.8796, "step": 615 }, { "epoch": 3.3776559287183003, "grad_norm": 0.1782795515564827, "learning_rate": 2.285229754387783e-05, "loss": 0.9047, "step": 616 }, { "epoch": 3.383139136394791, "grad_norm": 0.1808046435484797, "learning_rate": 2.2713802635596246e-05, "loss": 0.9014, "step": 617 }, { "epoch": 3.3886223440712815, "grad_norm": 0.16986006910284632, "learning_rate": 2.2575562076807857e-05, "loss": 0.9182, "step": 618 }, { "epoch": 3.3941055517477725, "grad_norm": 0.17065356826960065, "learning_rate": 2.2437577901587284e-05, "loss": 0.9148, "step": 619 }, { "epoch": 3.399588759424263, "grad_norm": 0.1821028903512155, "learning_rate": 2.22998521402368e-05, "loss": 0.892, "step": 620 }, { "epoch": 3.405071967100754, "grad_norm": 0.16756736884692283, "learning_rate": 2.216238681925628e-05, "loss": 0.9055, "step": 621 }, { "epoch": 3.4105551747772447, "grad_norm": 0.19298599105907382, "learning_rate": 2.2025183961313542e-05, "loss": 0.905, "step": 622 }, { "epoch": 3.4160383824537353, "grad_norm": 0.17599992995668898, "learning_rate": 2.188824558521452e-05, "loss": 0.9076, "step": 623 }, { "epoch": 3.4215215901302263, "grad_norm": 0.17170320698537844, "learning_rate": 2.175157370587348e-05, "loss": 0.9036, "step": 624 }, { "epoch": 3.427004797806717, "grad_norm": 0.19564360580394577, "learning_rate": 2.1615170334283535e-05, "loss": 0.8988, "step": 625 }, { "epoch": 3.432488005483208, "grad_norm": 0.15688766150962508, "learning_rate": 2.1479037477486936e-05, "loss": 0.8866, "step": 626 }, { "epoch": 3.4379712131596984, "grad_norm": 0.1814220481133191, "learning_rate": 2.1343177138545547e-05, "loss": 0.8957, "step": 627 }, { "epoch": 3.443454420836189, "grad_norm": 0.19751191092231474, "learning_rate": 2.1207591316511454e-05, "loss": 0.8952, "step": 628 }, { "epoch": 3.44893762851268, "grad_norm": 0.1646360237105222, "learning_rate": 2.1072282006397425e-05, "loss": 0.8953, "step": 629 }, { "epoch": 3.4544208361891706, "grad_norm": 0.18843073086840514, "learning_rate": 2.0937251199147684e-05, "loss": 0.9038, "step": 630 }, { "epoch": 3.459904043865661, "grad_norm": 0.17451033699060078, "learning_rate": 2.0802500881608557e-05, "loss": 0.9059, "step": 631 }, { "epoch": 3.465387251542152, "grad_norm": 0.15864010191777436, "learning_rate": 2.066803303649918e-05, "loss": 0.8736, "step": 632 }, { "epoch": 3.470870459218643, "grad_norm": 0.1745241432795982, "learning_rate": 2.0533849642382446e-05, "loss": 0.8966, "step": 633 }, { "epoch": 3.476353666895134, "grad_norm": 0.15563829064977594, "learning_rate": 2.039995267363581e-05, "loss": 0.9013, "step": 634 }, { "epoch": 3.4818368745716244, "grad_norm": 0.18758674942103779, "learning_rate": 2.026634410042223e-05, "loss": 0.8947, "step": 635 }, { "epoch": 3.487320082248115, "grad_norm": 0.15887193712810688, "learning_rate": 2.0133025888661263e-05, "loss": 0.9054, "step": 636 }, { "epoch": 3.492803289924606, "grad_norm": 0.17452023322549837, "learning_rate": 2.0000000000000012e-05, "loss": 0.9076, "step": 637 }, { "epoch": 3.4982864976010966, "grad_norm": 0.18330071982251422, "learning_rate": 1.986726839178438e-05, "loss": 0.9015, "step": 638 }, { "epoch": 3.5037697052775876, "grad_norm": 0.1575715551896094, "learning_rate": 1.9734833017030227e-05, "loss": 0.8813, "step": 639 }, { "epoch": 3.509252912954078, "grad_norm": 0.18878562836114246, "learning_rate": 1.9602695824394576e-05, "loss": 0.9005, "step": 640 }, { "epoch": 3.5147361206305687, "grad_norm": 0.15533387161156822, "learning_rate": 1.9470858758147036e-05, "loss": 0.8757, "step": 641 }, { "epoch": 3.5202193283070597, "grad_norm": 0.1630547237218842, "learning_rate": 1.933932375814114e-05, "loss": 0.9038, "step": 642 }, { "epoch": 3.5257025359835503, "grad_norm": 0.1712109811813745, "learning_rate": 1.9208092759785818e-05, "loss": 0.9049, "step": 643 }, { "epoch": 3.531185743660041, "grad_norm": 0.14587673483344632, "learning_rate": 1.9077167694016903e-05, "loss": 0.8985, "step": 644 }, { "epoch": 3.536668951336532, "grad_norm": 0.16414938500317564, "learning_rate": 1.8946550487268706e-05, "loss": 0.9068, "step": 645 }, { "epoch": 3.5421521590130225, "grad_norm": 0.15376745325776353, "learning_rate": 1.8816243061445734e-05, "loss": 0.8955, "step": 646 }, { "epoch": 3.5476353666895135, "grad_norm": 0.15960811991170182, "learning_rate": 1.8686247333894366e-05, "loss": 0.8967, "step": 647 }, { "epoch": 3.553118574366004, "grad_norm": 0.15855046926600352, "learning_rate": 1.8556565217374606e-05, "loss": 0.8951, "step": 648 }, { "epoch": 3.558601782042495, "grad_norm": 0.15624502487015215, "learning_rate": 1.8427198620032037e-05, "loss": 0.9061, "step": 649 }, { "epoch": 3.5640849897189857, "grad_norm": 0.1679249396514824, "learning_rate": 1.829814944536963e-05, "loss": 0.9094, "step": 650 }, { "epoch": 3.5695681973954763, "grad_norm": 0.15999412608591182, "learning_rate": 1.8169419592219813e-05, "loss": 0.8991, "step": 651 }, { "epoch": 3.5750514050719673, "grad_norm": 0.17340179382194568, "learning_rate": 1.8041010954716544e-05, "loss": 0.9103, "step": 652 }, { "epoch": 3.580534612748458, "grad_norm": 0.15237872236931496, "learning_rate": 1.7912925422267345e-05, "loss": 0.8999, "step": 653 }, { "epoch": 3.5860178204249484, "grad_norm": 1.2272413852202702, "learning_rate": 1.7785164879525604e-05, "loss": 0.9349, "step": 654 }, { "epoch": 3.5915010281014395, "grad_norm": 0.17476206012460227, "learning_rate": 1.7657731206362813e-05, "loss": 0.9132, "step": 655 }, { "epoch": 3.59698423577793, "grad_norm": 0.1762962782430207, "learning_rate": 1.7530626277840846e-05, "loss": 0.8938, "step": 656 }, { "epoch": 3.6024674434544206, "grad_norm": 0.181015770218541, "learning_rate": 1.7403851964184486e-05, "loss": 0.9069, "step": 657 }, { "epoch": 3.6079506511309116, "grad_norm": 0.16366717982789378, "learning_rate": 1.7277410130753775e-05, "loss": 0.8954, "step": 658 }, { "epoch": 3.613433858807402, "grad_norm": 0.1612334600367671, "learning_rate": 1.7151302638016683e-05, "loss": 0.884, "step": 659 }, { "epoch": 3.618917066483893, "grad_norm": 0.16685277983540153, "learning_rate": 1.7025531341521685e-05, "loss": 0.894, "step": 660 }, { "epoch": 3.624400274160384, "grad_norm": 0.1669827992942319, "learning_rate": 1.690009809187041e-05, "loss": 0.8952, "step": 661 }, { "epoch": 3.629883481836875, "grad_norm": 0.15253380478976128, "learning_rate": 1.6775004734690495e-05, "loss": 0.8864, "step": 662 }, { "epoch": 3.6353666895133654, "grad_norm": 0.15609996007733176, "learning_rate": 1.6650253110608415e-05, "loss": 0.9006, "step": 663 }, { "epoch": 3.640849897189856, "grad_norm": 0.1461455903080232, "learning_rate": 1.6525845055222306e-05, "loss": 0.917, "step": 664 }, { "epoch": 3.646333104866347, "grad_norm": 0.15716311122738527, "learning_rate": 1.6401782399075098e-05, "loss": 0.9149, "step": 665 }, { "epoch": 3.6518163125428376, "grad_norm": 0.32627646467352905, "learning_rate": 1.627806696762745e-05, "loss": 0.9054, "step": 666 }, { "epoch": 3.657299520219328, "grad_norm": 0.16854342857121438, "learning_rate": 1.615470058123099e-05, "loss": 0.8805, "step": 667 }, { "epoch": 3.662782727895819, "grad_norm": 0.1747226962814483, "learning_rate": 1.603168505510148e-05, "loss": 0.9052, "step": 668 }, { "epoch": 3.6682659355723097, "grad_norm": 0.16903191589677927, "learning_rate": 1.5909022199292104e-05, "loss": 0.8961, "step": 669 }, { "epoch": 3.6737491432488003, "grad_norm": 8.87107113952348, "learning_rate": 1.5786713818666876e-05, "loss": 0.9752, "step": 670 }, { "epoch": 3.6792323509252913, "grad_norm": 0.20674605140421184, "learning_rate": 1.566476171287401e-05, "loss": 0.8898, "step": 671 }, { "epoch": 3.684715558601782, "grad_norm": 0.19629370706239815, "learning_rate": 1.554316767631951e-05, "loss": 0.9059, "step": 672 }, { "epoch": 3.690198766278273, "grad_norm": 0.21259237974634135, "learning_rate": 1.5421933498140763e-05, "loss": 0.8958, "step": 673 }, { "epoch": 3.6956819739547635, "grad_norm": 0.18310426924162343, "learning_rate": 1.5301060962180133e-05, "loss": 0.8853, "step": 674 }, { "epoch": 3.7011651816312545, "grad_norm": 0.1751171847378099, "learning_rate": 1.518055184695884e-05, "loss": 0.8813, "step": 675 }, { "epoch": 3.706648389307745, "grad_norm": 0.16886538069007115, "learning_rate": 1.5060407925650662e-05, "loss": 0.8881, "step": 676 }, { "epoch": 3.7121315969842357, "grad_norm": 0.17046906947879148, "learning_rate": 1.494063096605595e-05, "loss": 0.8937, "step": 677 }, { "epoch": 3.7176148046607267, "grad_norm": 0.1581379052592717, "learning_rate": 1.4821222730575561e-05, "loss": 0.8972, "step": 678 }, { "epoch": 3.7230980123372173, "grad_norm": 0.6969777085406519, "learning_rate": 1.4702184976184915e-05, "loss": 0.9041, "step": 679 }, { "epoch": 3.728581220013708, "grad_norm": 0.2182887056206488, "learning_rate": 1.4583519454408191e-05, "loss": 0.9039, "step": 680 }, { "epoch": 3.734064427690199, "grad_norm": 0.22945666556054492, "learning_rate": 1.4465227911292537e-05, "loss": 0.8887, "step": 681 }, { "epoch": 3.7395476353666894, "grad_norm": 0.22736819225186297, "learning_rate": 1.434731208738232e-05, "loss": 0.9121, "step": 682 }, { "epoch": 3.74503084304318, "grad_norm": 0.20165874554977742, "learning_rate": 1.4229773717693625e-05, "loss": 0.8965, "step": 683 }, { "epoch": 3.750514050719671, "grad_norm": 0.184903868108441, "learning_rate": 1.4112614531688645e-05, "loss": 0.9045, "step": 684 }, { "epoch": 3.7559972583961616, "grad_norm": 0.18326386101695807, "learning_rate": 1.3995836253250233e-05, "loss": 0.8995, "step": 685 }, { "epoch": 3.7614804660726526, "grad_norm": 0.20107358682198542, "learning_rate": 1.3879440600656607e-05, "loss": 0.9072, "step": 686 }, { "epoch": 3.766963673749143, "grad_norm": 0.3569275389027143, "learning_rate": 1.3763429286555963e-05, "loss": 0.9048, "step": 687 }, { "epoch": 3.7724468814256342, "grad_norm": 0.19199246127012237, "learning_rate": 1.3647804017941373e-05, "loss": 0.8941, "step": 688 }, { "epoch": 3.777930089102125, "grad_norm": 0.19373559370826268, "learning_rate": 1.3532566496125634e-05, "loss": 0.9014, "step": 689 }, { "epoch": 3.7834132967786154, "grad_norm": 0.180911982737044, "learning_rate": 1.3417718416716183e-05, "loss": 0.8911, "step": 690 }, { "epoch": 3.7888965044551064, "grad_norm": 0.17759995896183428, "learning_rate": 1.3303261469590228e-05, "loss": 0.8932, "step": 691 }, { "epoch": 3.794379712131597, "grad_norm": 0.159075207881893, "learning_rate": 1.3189197338869853e-05, "loss": 0.882, "step": 692 }, { "epoch": 3.7998629198080875, "grad_norm": 0.18045574791107008, "learning_rate": 1.3075527702897185e-05, "loss": 0.9228, "step": 693 }, { "epoch": 3.8053461274845786, "grad_norm": 0.16316813300832433, "learning_rate": 1.2962254234209826e-05, "loss": 0.8894, "step": 694 }, { "epoch": 3.810829335161069, "grad_norm": 0.17660313697972807, "learning_rate": 1.2849378599516085e-05, "loss": 0.8997, "step": 695 }, { "epoch": 3.8163125428375597, "grad_norm": 0.1679766974660261, "learning_rate": 1.273690245967059e-05, "loss": 0.9042, "step": 696 }, { "epoch": 3.8217957505140507, "grad_norm": 0.15705812295975527, "learning_rate": 1.2624827469649739e-05, "loss": 0.8959, "step": 697 }, { "epoch": 3.8272789581905413, "grad_norm": 0.1600279121047265, "learning_rate": 1.2513155278527446e-05, "loss": 0.8942, "step": 698 }, { "epoch": 3.8327621658670323, "grad_norm": 0.14891745897684322, "learning_rate": 1.240188752945084e-05, "loss": 0.9133, "step": 699 }, { "epoch": 3.838245373543523, "grad_norm": 1.0164234645051264, "learning_rate": 1.2291025859616026e-05, "loss": 0.8963, "step": 700 }, { "epoch": 3.843728581220014, "grad_norm": 0.15287661592525834, "learning_rate": 1.21805719002441e-05, "loss": 0.896, "step": 701 }, { "epoch": 3.8492117888965045, "grad_norm": 0.18515627168326385, "learning_rate": 1.2070527276557092e-05, "loss": 0.8907, "step": 702 }, { "epoch": 3.854694996572995, "grad_norm": 0.19538289491293945, "learning_rate": 1.1960893607754022e-05, "loss": 0.8866, "step": 703 }, { "epoch": 3.860178204249486, "grad_norm": 0.1803420043271255, "learning_rate": 1.1851672506987165e-05, "loss": 0.9027, "step": 704 }, { "epoch": 3.8656614119259767, "grad_norm": 0.1728624061069401, "learning_rate": 1.17428655813382e-05, "loss": 0.9088, "step": 705 }, { "epoch": 3.8711446196024673, "grad_norm": 0.18740842748319184, "learning_rate": 1.1634474431794676e-05, "loss": 0.9012, "step": 706 }, { "epoch": 3.8766278272789583, "grad_norm": 0.16138769768933767, "learning_rate": 1.1526500653226385e-05, "loss": 0.9168, "step": 707 }, { "epoch": 3.882111034955449, "grad_norm": 0.14601780097035175, "learning_rate": 1.141894583436189e-05, "loss": 0.8923, "step": 708 }, { "epoch": 3.8875942426319394, "grad_norm": 0.16268490177886352, "learning_rate": 1.1311811557765208e-05, "loss": 0.9011, "step": 709 }, { "epoch": 3.8930774503084304, "grad_norm": 0.14407863627045317, "learning_rate": 1.1205099399812478e-05, "loss": 0.8937, "step": 710 }, { "epoch": 3.898560657984921, "grad_norm": 0.14837780514070958, "learning_rate": 1.1098810930668754e-05, "loss": 0.9035, "step": 711 }, { "epoch": 3.904043865661412, "grad_norm": 0.13890190728316182, "learning_rate": 1.0992947714264952e-05, "loss": 0.8925, "step": 712 }, { "epoch": 3.9095270733379026, "grad_norm": 0.1482520122684546, "learning_rate": 1.088751130827478e-05, "loss": 0.8864, "step": 713 }, { "epoch": 3.9150102810143936, "grad_norm": 0.19175152200646928, "learning_rate": 1.078250326409188e-05, "loss": 0.9022, "step": 714 }, { "epoch": 3.920493488690884, "grad_norm": 0.14176488398671347, "learning_rate": 1.0677925126806956e-05, "loss": 0.9017, "step": 715 }, { "epoch": 3.925976696367375, "grad_norm": 0.14891595552628026, "learning_rate": 1.0573778435185039e-05, "loss": 0.9038, "step": 716 }, { "epoch": 3.931459904043866, "grad_norm": 0.13751484111663212, "learning_rate": 1.047006472164287e-05, "loss": 0.8925, "step": 717 }, { "epoch": 3.9369431117203564, "grad_norm": 0.1431600770457188, "learning_rate": 1.0366785512226359e-05, "loss": 0.903, "step": 718 }, { "epoch": 3.942426319396847, "grad_norm": 0.7396147350976154, "learning_rate": 1.0263942326588054e-05, "loss": 0.9243, "step": 719 }, { "epoch": 3.947909527073338, "grad_norm": 0.18213078561266646, "learning_rate": 1.0161536677964933e-05, "loss": 0.8895, "step": 720 }, { "epoch": 3.9533927347498286, "grad_norm": 0.19698575071196872, "learning_rate": 1.0059570073155953e-05, "loss": 0.8988, "step": 721 }, { "epoch": 3.958875942426319, "grad_norm": 0.16625311791486364, "learning_rate": 9.958044012500023e-06, "loss": 0.8841, "step": 722 }, { "epoch": 3.96435915010281, "grad_norm": 0.18406574667380657, "learning_rate": 9.856959989853876e-06, "loss": 0.8959, "step": 723 }, { "epoch": 3.969842357779301, "grad_norm": 0.16041228708375482, "learning_rate": 9.75631949257004e-06, "loss": 0.8885, "step": 724 }, { "epoch": 3.9753255654557917, "grad_norm": 0.16790760471515281, "learning_rate": 9.656124001475068e-06, "loss": 0.8832, "step": 725 }, { "epoch": 3.9808087731322823, "grad_norm": 0.1643989573541531, "learning_rate": 9.556374990847618e-06, "loss": 0.9077, "step": 726 }, { "epoch": 3.9862919808087733, "grad_norm": 0.15089157844878578, "learning_rate": 9.457073928396871e-06, "loss": 0.8924, "step": 727 }, { "epoch": 3.991775188485264, "grad_norm": 0.15938422497231847, "learning_rate": 9.358222275240884e-06, "loss": 0.891, "step": 728 }, { "epoch": 3.9972583961617545, "grad_norm": 0.1386488462476611, "learning_rate": 9.25982148588506e-06, "loss": 0.8969, "step": 729 }, { "epoch": 4.0027416038382455, "grad_norm": 0.34700978060414556, "learning_rate": 9.161873008200816e-06, "loss": 1.4657, "step": 730 }, { "epoch": 4.0082248115147365, "grad_norm": 0.2465097801021323, "learning_rate": 9.064378283404247e-06, "loss": 0.8725, "step": 731 }, { "epoch": 4.013708019191227, "grad_norm": 0.26497344033852505, "learning_rate": 8.967338746034882e-06, "loss": 0.8759, "step": 732 }, { "epoch": 4.019191226867718, "grad_norm": 0.21838213857553232, "learning_rate": 8.870755823934662e-06, "loss": 0.8627, "step": 733 }, { "epoch": 4.024674434544209, "grad_norm": 0.20277610227352777, "learning_rate": 8.774630938226831e-06, "loss": 0.886, "step": 734 }, { "epoch": 4.030157642220699, "grad_norm": 0.24713492980404156, "learning_rate": 8.678965503295114e-06, "loss": 0.8692, "step": 735 }, { "epoch": 4.03564084989719, "grad_norm": 0.18596227337602494, "learning_rate": 8.583760926762852e-06, "loss": 0.8702, "step": 736 }, { "epoch": 4.041124057573681, "grad_norm": 0.2255598033578029, "learning_rate": 8.489018609472297e-06, "loss": 0.859, "step": 737 }, { "epoch": 4.046607265250171, "grad_norm": 0.19671824377342131, "learning_rate": 8.394739945464016e-06, "loss": 0.8825, "step": 738 }, { "epoch": 4.052090472926662, "grad_norm": 0.19536276470724584, "learning_rate": 8.300926321956391e-06, "loss": 0.8655, "step": 739 }, { "epoch": 4.057573680603153, "grad_norm": 0.20553044507457272, "learning_rate": 8.207579119325145e-06, "loss": 0.8581, "step": 740 }, { "epoch": 4.063056888279643, "grad_norm": 0.1745254415584832, "learning_rate": 8.114699711083113e-06, "loss": 0.8559, "step": 741 }, { "epoch": 4.068540095956134, "grad_norm": 0.19419389496880718, "learning_rate": 8.022289463859963e-06, "loss": 0.8684, "step": 742 }, { "epoch": 4.074023303632625, "grad_norm": 0.16866208377226335, "learning_rate": 7.930349737382137e-06, "loss": 0.8603, "step": 743 }, { "epoch": 4.079506511309116, "grad_norm": 0.17045009966171426, "learning_rate": 7.838881884452827e-06, "loss": 0.8775, "step": 744 }, { "epoch": 4.084989718985606, "grad_norm": 0.1672289163103852, "learning_rate": 7.747887250932047e-06, "loss": 0.8577, "step": 745 }, { "epoch": 4.090472926662097, "grad_norm": 0.16957620944180635, "learning_rate": 7.657367175716884e-06, "loss": 0.8573, "step": 746 }, { "epoch": 4.095956134338588, "grad_norm": 0.15814438359723273, "learning_rate": 7.5673229907217146e-06, "loss": 0.8519, "step": 747 }, { "epoch": 4.1014393420150785, "grad_norm": 0.1554296876299363, "learning_rate": 7.477756020858695e-06, "loss": 0.8635, "step": 748 }, { "epoch": 4.10692254969157, "grad_norm": 0.16052588316393465, "learning_rate": 7.38866758401823e-06, "loss": 0.8723, "step": 749 }, { "epoch": 4.112405757368061, "grad_norm": 0.15330364083531656, "learning_rate": 7.300058991049534e-06, "loss": 0.8828, "step": 750 }, { "epoch": 4.117888965044551, "grad_norm": 0.1495227789255928, "learning_rate": 7.211931545741433e-06, "loss": 0.8506, "step": 751 }, { "epoch": 4.123372172721042, "grad_norm": 0.15408046491976884, "learning_rate": 7.124286544803136e-06, "loss": 0.8668, "step": 752 }, { "epoch": 4.128855380397533, "grad_norm": 0.14881498638247898, "learning_rate": 7.037125277845112e-06, "loss": 0.8801, "step": 753 }, { "epoch": 4.134338588074023, "grad_norm": 0.1464102070171832, "learning_rate": 6.950449027360213e-06, "loss": 0.8726, "step": 754 }, { "epoch": 4.139821795750514, "grad_norm": 0.1453166333634366, "learning_rate": 6.864259068704688e-06, "loss": 0.8607, "step": 755 }, { "epoch": 4.145305003427005, "grad_norm": 0.13625555326942757, "learning_rate": 6.778556670079535e-06, "loss": 0.8792, "step": 756 }, { "epoch": 4.150788211103496, "grad_norm": 0.13807498863680387, "learning_rate": 6.69334309251175e-06, "loss": 0.8474, "step": 757 }, { "epoch": 4.156271418779986, "grad_norm": 0.14308904016946167, "learning_rate": 6.608619589835803e-06, "loss": 0.8683, "step": 758 }, { "epoch": 4.161754626456477, "grad_norm": 0.1440963907671081, "learning_rate": 6.524387408675208e-06, "loss": 0.8633, "step": 759 }, { "epoch": 4.167237834132968, "grad_norm": 0.14254526065135373, "learning_rate": 6.440647788424166e-06, "loss": 0.8714, "step": 760 }, { "epoch": 4.172721041809458, "grad_norm": 0.15400971348680637, "learning_rate": 6.357401961229293e-06, "loss": 0.8719, "step": 761 }, { "epoch": 4.178204249485949, "grad_norm": 0.1334821564613061, "learning_rate": 6.274651151971567e-06, "loss": 0.8604, "step": 762 }, { "epoch": 4.18368745716244, "grad_norm": 0.15458395585147408, "learning_rate": 6.1923965782482165e-06, "loss": 0.8855, "step": 763 }, { "epoch": 4.18917066483893, "grad_norm": 0.14422168418866493, "learning_rate": 6.110639450354882e-06, "loss": 0.8709, "step": 764 }, { "epoch": 4.194653872515421, "grad_norm": 0.13640204014715013, "learning_rate": 6.0293809712677775e-06, "loss": 0.8779, "step": 765 }, { "epoch": 4.2001370801919125, "grad_norm": 0.13728381458929773, "learning_rate": 5.9486223366259555e-06, "loss": 0.863, "step": 766 }, { "epoch": 4.205620287868403, "grad_norm": 1.0957146408038854, "learning_rate": 5.868364734713776e-06, "loss": 0.8706, "step": 767 }, { "epoch": 4.211103495544894, "grad_norm": 0.14870058965626423, "learning_rate": 5.788609346443386e-06, "loss": 0.8808, "step": 768 }, { "epoch": 4.216586703221385, "grad_norm": 0.16703761517821936, "learning_rate": 5.70935734533733e-06, "loss": 0.852, "step": 769 }, { "epoch": 4.222069910897876, "grad_norm": 0.1844114299522098, "learning_rate": 5.630609897511328e-06, "loss": 0.8613, "step": 770 }, { "epoch": 4.227553118574366, "grad_norm": 0.17398006404372696, "learning_rate": 5.552368161657082e-06, "loss": 0.8556, "step": 771 }, { "epoch": 4.233036326250857, "grad_norm": 0.16676512516292483, "learning_rate": 5.474633289025244e-06, "loss": 0.8425, "step": 772 }, { "epoch": 4.238519533927348, "grad_norm": 0.15690331804763263, "learning_rate": 5.397406423408446e-06, "loss": 0.874, "step": 773 }, { "epoch": 4.244002741603838, "grad_norm": 0.15751054116710525, "learning_rate": 5.3206887011245165e-06, "loss": 0.8871, "step": 774 }, { "epoch": 4.249485949280329, "grad_norm": 0.1535880170728185, "learning_rate": 5.24448125099974e-06, "loss": 0.8681, "step": 775 }, { "epoch": 4.25496915695682, "grad_norm": 0.15063251461992208, "learning_rate": 5.1687851943522215e-06, "loss": 0.8744, "step": 776 }, { "epoch": 4.26045236463331, "grad_norm": 0.1520419945091867, "learning_rate": 5.093601644975428e-06, "loss": 0.8681, "step": 777 }, { "epoch": 4.265935572309801, "grad_norm": 0.14153569721366746, "learning_rate": 5.018931709121791e-06, "loss": 0.8638, "step": 778 }, { "epoch": 4.271418779986292, "grad_norm": 0.14202440325296312, "learning_rate": 4.9447764854863915e-06, "loss": 0.8595, "step": 779 }, { "epoch": 4.276901987662782, "grad_norm": 0.13789016773017412, "learning_rate": 4.871137065190854e-06, "loss": 0.8738, "step": 780 }, { "epoch": 4.282385195339273, "grad_norm": 0.13587353525443027, "learning_rate": 4.798014531767261e-06, "loss": 0.8665, "step": 781 }, { "epoch": 4.287868403015764, "grad_norm": 0.13265718560401082, "learning_rate": 4.725409961142173e-06, "loss": 0.874, "step": 782 }, { "epoch": 4.293351610692255, "grad_norm": 0.13693830134507576, "learning_rate": 4.653324421620884e-06, "loss": 0.8619, "step": 783 }, { "epoch": 4.2988348183687455, "grad_norm": 0.13146748198275093, "learning_rate": 4.581758973871609e-06, "loss": 0.8825, "step": 784 }, { "epoch": 4.3043180260452365, "grad_norm": 0.13405773370704815, "learning_rate": 4.510714670909946e-06, "loss": 0.8536, "step": 785 }, { "epoch": 4.3098012337217275, "grad_norm": 0.14574117129879566, "learning_rate": 4.440192558083367e-06, "loss": 0.8506, "step": 786 }, { "epoch": 4.315284441398218, "grad_norm": 0.13101655607771853, "learning_rate": 4.370193673055787e-06, "loss": 0.8636, "step": 787 }, { "epoch": 4.320767649074709, "grad_norm": 0.13191753249596408, "learning_rate": 4.300719045792376e-06, "loss": 0.8671, "step": 788 }, { "epoch": 4.3262508567512, "grad_norm": 0.1310909838553508, "learning_rate": 4.231769698544352e-06, "loss": 0.8924, "step": 789 }, { "epoch": 4.33173406442769, "grad_norm": 0.13210093237615378, "learning_rate": 4.163346645833928e-06, "loss": 0.8605, "step": 790 }, { "epoch": 4.337217272104181, "grad_norm": 0.128792479044049, "learning_rate": 4.0954508944394474e-06, "loss": 0.8695, "step": 791 }, { "epoch": 4.342700479780672, "grad_norm": 0.22455677717587358, "learning_rate": 4.028083443380486e-06, "loss": 0.8691, "step": 792 }, { "epoch": 4.348183687457162, "grad_norm": 0.40527297890394276, "learning_rate": 3.961245283903239e-06, "loss": 0.8805, "step": 793 }, { "epoch": 4.353666895133653, "grad_norm": 0.2888577519432354, "learning_rate": 3.89493739946587e-06, "loss": 0.8803, "step": 794 }, { "epoch": 4.359150102810144, "grad_norm": 0.1350998279076715, "learning_rate": 3.829160765724052e-06, "loss": 0.8513, "step": 795 }, { "epoch": 4.364633310486635, "grad_norm": 0.1309943366405807, "learning_rate": 3.7639163505166633e-06, "loss": 0.8668, "step": 796 }, { "epoch": 4.370116518163125, "grad_norm": 0.12953103571452862, "learning_rate": 3.6992051138514717e-06, "loss": 0.8712, "step": 797 }, { "epoch": 4.375599725839616, "grad_norm": 0.13057599590166327, "learning_rate": 3.635028007891048e-06, "loss": 0.8686, "step": 798 }, { "epoch": 4.381082933516107, "grad_norm": 0.12973353273093458, "learning_rate": 3.5713859769387795e-06, "loss": 0.8831, "step": 799 }, { "epoch": 4.386566141192597, "grad_norm": 0.13131456969773786, "learning_rate": 3.5082799574249094e-06, "loss": 0.8639, "step": 800 }, { "epoch": 4.392049348869088, "grad_norm": 0.12595490195849035, "learning_rate": 3.4457108778928272e-06, "loss": 0.8631, "step": 801 }, { "epoch": 4.397532556545579, "grad_norm": 0.12813268032502836, "learning_rate": 3.3836796589853484e-06, "loss": 0.8663, "step": 802 }, { "epoch": 4.4030157642220695, "grad_norm": 0.12716051374477083, "learning_rate": 3.3221872134312184e-06, "loss": 0.8525, "step": 803 }, { "epoch": 4.4084989718985605, "grad_norm": 0.12371044909708838, "learning_rate": 3.261234446031658e-06, "loss": 0.8679, "step": 804 }, { "epoch": 4.413982179575052, "grad_norm": 0.12478585102448334, "learning_rate": 3.200822253647031e-06, "loss": 0.8889, "step": 805 }, { "epoch": 4.419465387251542, "grad_norm": 0.1293948457179513, "learning_rate": 3.140951525183691e-06, "loss": 0.8629, "step": 806 }, { "epoch": 4.424948594928033, "grad_norm": 0.127763766689417, "learning_rate": 3.0816231415808785e-06, "loss": 0.8545, "step": 807 }, { "epoch": 4.430431802604524, "grad_norm": 0.12507739335573223, "learning_rate": 3.02283797579773e-06, "loss": 0.8719, "step": 808 }, { "epoch": 4.435915010281015, "grad_norm": 0.12408679274948532, "learning_rate": 2.9645968928005085e-06, "loss": 0.8643, "step": 809 }, { "epoch": 4.441398217957505, "grad_norm": 0.14843016184470606, "learning_rate": 2.906900749549784e-06, "loss": 0.8993, "step": 810 }, { "epoch": 4.446881425633996, "grad_norm": 0.12190268058344725, "learning_rate": 2.849750394987907e-06, "loss": 0.8632, "step": 811 }, { "epoch": 4.452364633310487, "grad_norm": 0.12079449899424143, "learning_rate": 2.793146670026472e-06, "loss": 0.8792, "step": 812 }, { "epoch": 4.457847840986977, "grad_norm": 0.12301325559585843, "learning_rate": 2.737090407533938e-06, "loss": 0.8561, "step": 813 }, { "epoch": 4.463331048663468, "grad_norm": 0.12435308059508353, "learning_rate": 2.681582432323406e-06, "loss": 0.867, "step": 814 }, { "epoch": 4.468814256339959, "grad_norm": 0.12288901600568765, "learning_rate": 2.6266235611404645e-06, "loss": 0.8742, "step": 815 }, { "epoch": 4.474297464016449, "grad_norm": 0.12223916369585316, "learning_rate": 2.5722146026511574e-06, "loss": 0.8566, "step": 816 }, { "epoch": 4.47978067169294, "grad_norm": 0.12065119171288469, "learning_rate": 2.5183563574301185e-06, "loss": 0.869, "step": 817 }, { "epoch": 4.485263879369431, "grad_norm": 0.12212414616793427, "learning_rate": 2.465049617948778e-06, "loss": 0.8776, "step": 818 }, { "epoch": 4.490747087045921, "grad_norm": 0.12172913149868783, "learning_rate": 2.4122951685636674e-06, "loss": 0.8778, "step": 819 }, { "epoch": 4.496230294722412, "grad_norm": 0.12399214672845997, "learning_rate": 2.3600937855049467e-06, "loss": 0.872, "step": 820 }, { "epoch": 4.501713502398903, "grad_norm": 0.1209018345998114, "learning_rate": 2.308446236864916e-06, "loss": 0.8586, "step": 821 }, { "epoch": 4.5071967100753945, "grad_norm": 0.1177513204148733, "learning_rate": 2.257353282586774e-06, "loss": 0.8404, "step": 822 }, { "epoch": 4.512679917751885, "grad_norm": 0.12365836386703247, "learning_rate": 2.206815674453373e-06, "loss": 0.876, "step": 823 }, { "epoch": 4.518163125428376, "grad_norm": 0.12060944467881936, "learning_rate": 2.1568341560762152e-06, "loss": 0.8756, "step": 824 }, { "epoch": 4.523646333104867, "grad_norm": 0.1205949737711182, "learning_rate": 2.1074094628844754e-06, "loss": 0.8777, "step": 825 }, { "epoch": 4.529129540781357, "grad_norm": 0.12079575476686041, "learning_rate": 2.0585423221141807e-06, "loss": 0.8736, "step": 826 }, { "epoch": 4.534612748457848, "grad_norm": 0.1218689347455744, "learning_rate": 2.010233452797534e-06, "loss": 0.8667, "step": 827 }, { "epoch": 4.540095956134339, "grad_norm": 0.12283214110808495, "learning_rate": 1.9624835657523222e-06, "loss": 0.8666, "step": 828 }, { "epoch": 4.54557916381083, "grad_norm": 0.11913304460847574, "learning_rate": 1.9152933635714354e-06, "loss": 0.8667, "step": 829 }, { "epoch": 4.55106237148732, "grad_norm": 0.12377547685145791, "learning_rate": 1.8686635406125697e-06, "loss": 0.8898, "step": 830 }, { "epoch": 4.556545579163811, "grad_norm": 0.11633012806752614, "learning_rate": 1.822594782987972e-06, "loss": 0.8411, "step": 831 }, { "epoch": 4.562028786840301, "grad_norm": 0.11864592533450119, "learning_rate": 1.7770877685543687e-06, "loss": 0.8679, "step": 832 }, { "epoch": 4.567511994516792, "grad_norm": 0.12190861480595079, "learning_rate": 1.7321431669029953e-06, "loss": 0.8708, "step": 833 }, { "epoch": 4.572995202193283, "grad_norm": 0.12195430153778734, "learning_rate": 1.6877616393497075e-06, "loss": 0.8762, "step": 834 }, { "epoch": 4.578478409869774, "grad_norm": 0.12318104357609574, "learning_rate": 1.6439438389252948e-06, "loss": 0.8817, "step": 835 }, { "epoch": 4.583961617546264, "grad_norm": 0.1252397305163344, "learning_rate": 1.6006904103658572e-06, "loss": 0.8713, "step": 836 }, { "epoch": 4.589444825222755, "grad_norm": 0.11863485627738708, "learning_rate": 1.5580019901032929e-06, "loss": 0.8585, "step": 837 }, { "epoch": 4.594928032899246, "grad_norm": 0.11926883785074256, "learning_rate": 1.5158792062559813e-06, "loss": 0.8553, "step": 838 }, { "epoch": 4.6004112405757365, "grad_norm": 0.12010806317107317, "learning_rate": 1.4743226786194931e-06, "loss": 0.8632, "step": 839 }, { "epoch": 4.6058944482522275, "grad_norm": 0.1209497918934812, "learning_rate": 1.4333330186575079e-06, "loss": 0.8661, "step": 840 }, { "epoch": 4.6113776559287185, "grad_norm": 0.12015700387003853, "learning_rate": 1.3929108294927951e-06, "loss": 0.8672, "step": 841 }, { "epoch": 4.6168608636052095, "grad_norm": 0.12422998319235647, "learning_rate": 1.3530567058983369e-06, "loss": 0.8732, "step": 842 }, { "epoch": 4.6223440712817, "grad_norm": 0.11888832603251932, "learning_rate": 1.31377123428861e-06, "loss": 0.8723, "step": 843 }, { "epoch": 4.627827278958191, "grad_norm": 0.11996366586250991, "learning_rate": 1.2750549927109136e-06, "loss": 0.8574, "step": 844 }, { "epoch": 4.633310486634681, "grad_norm": 0.11745160261296496, "learning_rate": 1.2369085508368862e-06, "loss": 0.8604, "step": 845 }, { "epoch": 4.638793694311172, "grad_norm": 0.12200784232729492, "learning_rate": 1.1993324699541265e-06, "loss": 0.8517, "step": 846 }, { "epoch": 4.644276901987663, "grad_norm": 0.11775847148513344, "learning_rate": 1.1623273029579195e-06, "loss": 0.8695, "step": 847 }, { "epoch": 4.649760109664154, "grad_norm": 0.11881330238355213, "learning_rate": 1.1258935943431237e-06, "loss": 0.8667, "step": 848 }, { "epoch": 4.655243317340644, "grad_norm": 0.11930205382890313, "learning_rate": 1.090031880196145e-06, "loss": 0.8669, "step": 849 }, { "epoch": 4.660726525017135, "grad_norm": 0.12062456998925507, "learning_rate": 1.0547426881870292e-06, "loss": 0.8706, "step": 850 }, { "epoch": 4.666209732693626, "grad_norm": 0.119609080366811, "learning_rate": 1.0200265375617514e-06, "loss": 0.8735, "step": 851 }, { "epoch": 4.671692940370116, "grad_norm": 0.11883822127861333, "learning_rate": 9.858839391345065e-07, "loss": 0.8857, "step": 852 }, { "epoch": 4.677176148046607, "grad_norm": 0.11869096130082298, "learning_rate": 9.523153952802633e-07, "loss": 0.8567, "step": 853 }, { "epoch": 4.682659355723098, "grad_norm": 0.11660180847816304, "learning_rate": 9.193213999273199e-07, "loss": 0.8732, "step": 854 }, { "epoch": 4.688142563399589, "grad_norm": 0.1156650459781897, "learning_rate": 8.869024385500524e-07, "loss": 0.8694, "step": 855 }, { "epoch": 4.693625771076079, "grad_norm": 0.11849635720653122, "learning_rate": 8.550589881617877e-07, "loss": 0.8666, "step": 856 }, { "epoch": 4.69910897875257, "grad_norm": 0.12360260427554048, "learning_rate": 8.237915173077681e-07, "loss": 0.8503, "step": 857 }, { "epoch": 4.7045921864290605, "grad_norm": 0.11834338075827068, "learning_rate": 7.93100486058247e-07, "loss": 0.8824, "step": 858 }, { "epoch": 4.7100753941055515, "grad_norm": 0.12094889766224019, "learning_rate": 7.629863460017506e-07, "loss": 0.8686, "step": 859 }, { "epoch": 4.715558601782043, "grad_norm": 0.12037069037280847, "learning_rate": 7.334495402383957e-07, "loss": 0.8668, "step": 860 }, { "epoch": 4.721041809458534, "grad_norm": 0.1165471688624166, "learning_rate": 7.044905033734096e-07, "loss": 0.878, "step": 861 }, { "epoch": 4.726525017135024, "grad_norm": 0.11868226704814998, "learning_rate": 6.761096615107043e-07, "loss": 0.8786, "step": 862 }, { "epoch": 4.732008224811515, "grad_norm": 0.11761938779683033, "learning_rate": 6.483074322466154e-07, "loss": 0.8801, "step": 863 }, { "epoch": 4.737491432488006, "grad_norm": 0.119444219753523, "learning_rate": 6.210842246637683e-07, "loss": 0.8734, "step": 864 }, { "epoch": 4.742974640164496, "grad_norm": 0.11552293652107705, "learning_rate": 5.944404393250481e-07, "loss": 0.8443, "step": 865 }, { "epoch": 4.748457847840987, "grad_norm": 0.11624067828548895, "learning_rate": 5.683764682677018e-07, "loss": 0.8826, "step": 866 }, { "epoch": 4.753941055517478, "grad_norm": 0.11747660823290842, "learning_rate": 5.428926949975788e-07, "loss": 0.871, "step": 867 }, { "epoch": 4.759424263193969, "grad_norm": 0.11800804895284038, "learning_rate": 5.179894944834863e-07, "loss": 0.8808, "step": 868 }, { "epoch": 4.764907470870459, "grad_norm": 0.11426381832520421, "learning_rate": 4.936672331516778e-07, "loss": 0.865, "step": 869 }, { "epoch": 4.77039067854695, "grad_norm": 0.11934453977172532, "learning_rate": 4.699262688804451e-07, "loss": 0.8476, "step": 870 }, { "epoch": 4.77587388622344, "grad_norm": 0.11932772884879854, "learning_rate": 4.467669509948591e-07, "loss": 0.8644, "step": 871 }, { "epoch": 4.781357093899931, "grad_norm": 0.11533502002121204, "learning_rate": 4.241896202616502e-07, "loss": 0.8659, "step": 872 }, { "epoch": 4.786840301576422, "grad_norm": 0.11707563398942186, "learning_rate": 4.0219460888415884e-07, "loss": 0.8687, "step": 873 }, { "epoch": 4.792323509252913, "grad_norm": 0.11572299884307491, "learning_rate": 3.807822404974726e-07, "loss": 0.8619, "step": 874 }, { "epoch": 4.797806716929403, "grad_norm": 0.11414580388031612, "learning_rate": 3.599528301636612e-07, "loss": 0.8773, "step": 875 }, { "epoch": 4.803289924605894, "grad_norm": 0.1139811437838818, "learning_rate": 3.397066843671315e-07, "loss": 0.8614, "step": 876 }, { "epoch": 4.8087731322823855, "grad_norm": 0.11506534996205621, "learning_rate": 3.200441010101196e-07, "loss": 0.851, "step": 877 }, { "epoch": 4.814256339958876, "grad_norm": 0.11489034470022166, "learning_rate": 3.0096536940832145e-07, "loss": 0.8847, "step": 878 }, { "epoch": 4.819739547635367, "grad_norm": 0.11675777776548621, "learning_rate": 2.824707702866114e-07, "loss": 0.8569, "step": 879 }, { "epoch": 4.825222755311858, "grad_norm": 0.11463623776794303, "learning_rate": 2.645605757749392e-07, "loss": 0.8607, "step": 880 }, { "epoch": 4.830705962988349, "grad_norm": 0.11596024436355572, "learning_rate": 2.4723504940430187e-07, "loss": 0.8815, "step": 881 }, { "epoch": 4.836189170664839, "grad_norm": 0.11832403238695531, "learning_rate": 2.3049444610288462e-07, "loss": 0.8729, "step": 882 }, { "epoch": 4.84167237834133, "grad_norm": 0.1163599327714919, "learning_rate": 2.1433901219229502e-07, "loss": 0.8607, "step": 883 }, { "epoch": 4.84715558601782, "grad_norm": 0.11506741204644842, "learning_rate": 1.9876898538394362e-07, "loss": 0.875, "step": 884 }, { "epoch": 4.852638793694311, "grad_norm": 0.11571697024128855, "learning_rate": 1.8378459477555788e-07, "loss": 0.876, "step": 885 }, { "epoch": 4.858122001370802, "grad_norm": 0.1233092898569507, "learning_rate": 1.6938606084779375e-07, "loss": 0.8612, "step": 886 }, { "epoch": 4.863605209047293, "grad_norm": 0.11387876698076485, "learning_rate": 1.555735954610027e-07, "loss": 0.8499, "step": 887 }, { "epoch": 4.869088416723783, "grad_norm": 0.11498145411665506, "learning_rate": 1.4234740185210095e-07, "loss": 0.859, "step": 888 }, { "epoch": 4.874571624400274, "grad_norm": 0.11650609203788179, "learning_rate": 1.2970767463160284e-07, "loss": 0.8602, "step": 889 }, { "epoch": 4.880054832076765, "grad_norm": 0.11664236567167152, "learning_rate": 1.176545997807299e-07, "loss": 0.8746, "step": 890 }, { "epoch": 4.885538039753255, "grad_norm": 0.11702468705401428, "learning_rate": 1.0618835464870191e-07, "loss": 0.8525, "step": 891 }, { "epoch": 4.891021247429746, "grad_norm": 0.11451498374169765, "learning_rate": 9.530910795009895e-08, "loss": 0.8599, "step": 892 }, { "epoch": 4.896504455106237, "grad_norm": 0.11475845584891285, "learning_rate": 8.501701976239673e-08, "loss": 0.8583, "step": 893 }, { "epoch": 4.901987662782728, "grad_norm": 0.11547791831787842, "learning_rate": 7.531224152362183e-08, "loss": 0.8655, "step": 894 }, { "epoch": 4.9074708704592185, "grad_norm": 0.11917262127195702, "learning_rate": 6.619491603008676e-08, "loss": 0.8724, "step": 895 }, { "epoch": 4.9129540781357095, "grad_norm": 0.11681023683693044, "learning_rate": 5.766517743432953e-08, "loss": 0.8638, "step": 896 }, { "epoch": 4.9184372858122005, "grad_norm": 0.11279645464293689, "learning_rate": 4.9723151243106225e-08, "loss": 0.8528, "step": 897 }, { "epoch": 4.923920493488691, "grad_norm": 0.1138256632587849, "learning_rate": 4.236895431557031e-08, "loss": 0.8572, "step": 898 }, { "epoch": 4.929403701165182, "grad_norm": 0.11375846263511914, "learning_rate": 3.560269486154066e-08, "loss": 0.8391, "step": 899 }, { "epoch": 4.934886908841673, "grad_norm": 0.11717016335228325, "learning_rate": 2.9424472439911757e-08, "loss": 0.8752, "step": 900 }, { "epoch": 4.940370116518163, "grad_norm": 0.11269596219934125, "learning_rate": 2.3834377957183684e-08, "loss": 0.8491, "step": 901 }, { "epoch": 4.945853324194654, "grad_norm": 0.1158772425437692, "learning_rate": 1.8832493666125494e-08, "loss": 0.872, "step": 902 }, { "epoch": 4.951336531871145, "grad_norm": 0.9980363287282, "learning_rate": 1.4418893164585002e-08, "loss": 0.9242, "step": 903 }, { "epoch": 4.956819739547635, "grad_norm": 0.11610912640066366, "learning_rate": 1.0593641394369691e-08, "loss": 0.8648, "step": 904 }, { "epoch": 4.962302947224126, "grad_norm": 0.11625433625052747, "learning_rate": 7.356794640318576e-09, "loss": 0.857, "step": 905 }, { "epoch": 4.967786154900617, "grad_norm": 0.11721935619429466, "learning_rate": 4.708400529476187e-09, "loss": 0.85, "step": 906 }, { "epoch": 4.973269362577108, "grad_norm": 0.11503586503967776, "learning_rate": 2.648498030364266e-09, "loss": 0.8709, "step": 907 }, { "epoch": 4.978752570253598, "grad_norm": 0.11671324103566125, "learning_rate": 1.1771174524355388e-09, "loss": 0.8843, "step": 908 }, { "epoch": 4.984235777930089, "grad_norm": 0.11519765971232111, "learning_rate": 2.9428044562074265e-10, "loss": 0.862, "step": 909 }, { "epoch": 4.98971898560658, "grad_norm": 0.11510258089520117, "learning_rate": 0.0, "loss": 0.8606, "step": 910 }, { "epoch": 4.98971898560658, "step": 910, "total_flos": 2.1065883137322516e+19, "train_loss": 1.0237522615836216, "train_runtime": 109255.8386, "train_samples_per_second": 4.272, "train_steps_per_second": 0.008 } ], "logging_steps": 1.0, "max_steps": 910, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.1065883137322516e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }