{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.0, "eval_steps": 500, "global_step": 5972, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0006697923643670462, "grad_norm": 56.23729480038479, "learning_rate": 8.361204013377928e-09, "loss": 3.8663, "step": 1 }, { "epoch": 0.0013395847287340924, "grad_norm": 47.345797880014125, "learning_rate": 1.6722408026755855e-08, "loss": 3.493, "step": 2 }, { "epoch": 0.0020093770931011385, "grad_norm": 47.15036342894708, "learning_rate": 2.508361204013378e-08, "loss": 3.4054, "step": 3 }, { "epoch": 0.0026791694574681848, "grad_norm": 56.35000038077596, "learning_rate": 3.344481605351171e-08, "loss": 3.6879, "step": 4 }, { "epoch": 0.003348961821835231, "grad_norm": 58.2104365887315, "learning_rate": 4.180602006688963e-08, "loss": 3.8914, "step": 5 }, { "epoch": 0.004018754186202277, "grad_norm": 45.94779387422799, "learning_rate": 5.016722408026756e-08, "loss": 3.2083, "step": 6 }, { "epoch": 0.004688546550569324, "grad_norm": 53.99451922851964, "learning_rate": 5.8528428093645485e-08, "loss": 3.8653, "step": 7 }, { "epoch": 0.0053583389149363695, "grad_norm": 44.13242268620964, "learning_rate": 6.688963210702342e-08, "loss": 3.6169, "step": 8 }, { "epoch": 0.006028131279303416, "grad_norm": 50.155530780142264, "learning_rate": 7.525083612040134e-08, "loss": 3.5221, "step": 9 }, { "epoch": 0.006697923643670462, "grad_norm": 53.315509543557226, "learning_rate": 8.361204013377927e-08, "loss": 3.6455, "step": 10 }, { "epoch": 0.007367716008037508, "grad_norm": 50.43111601747768, "learning_rate": 9.19732441471572e-08, "loss": 3.4714, "step": 11 }, { "epoch": 0.008037508372404554, "grad_norm": 50.211295185610496, "learning_rate": 1.0033444816053512e-07, "loss": 3.6226, "step": 12 }, { "epoch": 0.008707300736771601, "grad_norm": 54.44919594277917, "learning_rate": 1.0869565217391305e-07, "loss": 3.7747, "step": 13 }, { "epoch": 0.009377093101138647, "grad_norm": 49.860248321550344, "learning_rate": 1.1705685618729097e-07, "loss": 3.6608, "step": 14 }, { "epoch": 0.010046885465505693, "grad_norm": 52.88006048587579, "learning_rate": 1.254180602006689e-07, "loss": 3.6398, "step": 15 }, { "epoch": 0.010716677829872739, "grad_norm": 49.267007802567676, "learning_rate": 1.3377926421404684e-07, "loss": 3.4715, "step": 16 }, { "epoch": 0.011386470194239785, "grad_norm": 49.768727119823275, "learning_rate": 1.4214046822742475e-07, "loss": 3.3863, "step": 17 }, { "epoch": 0.012056262558606833, "grad_norm": 52.68633193618413, "learning_rate": 1.505016722408027e-07, "loss": 3.6426, "step": 18 }, { "epoch": 0.012726054922973878, "grad_norm": 54.606780754003786, "learning_rate": 1.5886287625418062e-07, "loss": 3.8026, "step": 19 }, { "epoch": 0.013395847287340924, "grad_norm": 42.82368348647926, "learning_rate": 1.6722408026755853e-07, "loss": 3.3452, "step": 20 }, { "epoch": 0.01406563965170797, "grad_norm": 51.87288797837393, "learning_rate": 1.7558528428093647e-07, "loss": 3.6638, "step": 21 }, { "epoch": 0.014735432016075016, "grad_norm": 51.86372369067128, "learning_rate": 1.839464882943144e-07, "loss": 3.5488, "step": 22 }, { "epoch": 0.015405224380442064, "grad_norm": 49.25551523413513, "learning_rate": 1.9230769230769234e-07, "loss": 3.3462, "step": 23 }, { "epoch": 0.016075016744809108, "grad_norm": 49.577893200106054, "learning_rate": 2.0066889632107025e-07, "loss": 3.5824, "step": 24 }, { "epoch": 0.016744809109176157, "grad_norm": 81.81428713969231, "learning_rate": 2.0903010033444818e-07, "loss": 3.4531, "step": 25 }, { "epoch": 0.017414601473543203, "grad_norm": 50.81267206287028, "learning_rate": 2.173913043478261e-07, "loss": 3.5673, "step": 26 }, { "epoch": 0.01808439383791025, "grad_norm": 47.943639934612804, "learning_rate": 2.2575250836120403e-07, "loss": 3.5453, "step": 27 }, { "epoch": 0.018754186202277295, "grad_norm": 49.32199333644242, "learning_rate": 2.3411371237458194e-07, "loss": 3.6768, "step": 28 }, { "epoch": 0.01942397856664434, "grad_norm": 47.701356229161384, "learning_rate": 2.424749163879599e-07, "loss": 3.6865, "step": 29 }, { "epoch": 0.020093770931011386, "grad_norm": 40.54863954172109, "learning_rate": 2.508361204013378e-07, "loss": 3.2281, "step": 30 }, { "epoch": 0.020763563295378432, "grad_norm": 43.04735920821606, "learning_rate": 2.591973244147157e-07, "loss": 3.2992, "step": 31 }, { "epoch": 0.021433355659745478, "grad_norm": 33.59697015837651, "learning_rate": 2.675585284280937e-07, "loss": 2.9454, "step": 32 }, { "epoch": 0.022103148024112524, "grad_norm": 46.636493813455516, "learning_rate": 2.759197324414716e-07, "loss": 3.2501, "step": 33 }, { "epoch": 0.02277294038847957, "grad_norm": 33.83194934167284, "learning_rate": 2.842809364548495e-07, "loss": 3.2977, "step": 34 }, { "epoch": 0.02344273275284662, "grad_norm": 35.645322347132996, "learning_rate": 2.9264214046822746e-07, "loss": 3.2548, "step": 35 }, { "epoch": 0.024112525117213665, "grad_norm": 33.52660742902596, "learning_rate": 3.010033444816054e-07, "loss": 2.9801, "step": 36 }, { "epoch": 0.02478231748158071, "grad_norm": 34.91778137895112, "learning_rate": 3.093645484949833e-07, "loss": 3.0003, "step": 37 }, { "epoch": 0.025452109845947757, "grad_norm": 27.526431204864792, "learning_rate": 3.1772575250836125e-07, "loss": 2.9559, "step": 38 }, { "epoch": 0.026121902210314803, "grad_norm": 34.552643269650645, "learning_rate": 3.2608695652173915e-07, "loss": 3.3614, "step": 39 }, { "epoch": 0.02679169457468185, "grad_norm": 34.41911615056053, "learning_rate": 3.3444816053511706e-07, "loss": 3.0677, "step": 40 }, { "epoch": 0.027461486939048894, "grad_norm": 36.02644901004904, "learning_rate": 3.42809364548495e-07, "loss": 3.0317, "step": 41 }, { "epoch": 0.02813127930341594, "grad_norm": 33.33430745333179, "learning_rate": 3.5117056856187294e-07, "loss": 3.0194, "step": 42 }, { "epoch": 0.028801071667782986, "grad_norm": 57.399494646276196, "learning_rate": 3.5953177257525085e-07, "loss": 3.4599, "step": 43 }, { "epoch": 0.029470864032150032, "grad_norm": 23.1757222540054, "learning_rate": 3.678929765886288e-07, "loss": 2.8285, "step": 44 }, { "epoch": 0.03014065639651708, "grad_norm": 26.795755544910506, "learning_rate": 3.7625418060200677e-07, "loss": 2.9764, "step": 45 }, { "epoch": 0.030810448760884127, "grad_norm": 29.131631587648414, "learning_rate": 3.846153846153847e-07, "loss": 3.2711, "step": 46 }, { "epoch": 0.03148024112525117, "grad_norm": 17.1347468019123, "learning_rate": 3.929765886287626e-07, "loss": 3.0186, "step": 47 }, { "epoch": 0.032150033489618215, "grad_norm": 15.551024735010813, "learning_rate": 4.013377926421405e-07, "loss": 2.7941, "step": 48 }, { "epoch": 0.03281982585398526, "grad_norm": 12.670095504905733, "learning_rate": 4.096989966555184e-07, "loss": 2.7342, "step": 49 }, { "epoch": 0.033489618218352314, "grad_norm": 12.308705958381482, "learning_rate": 4.1806020066889637e-07, "loss": 2.5127, "step": 50 }, { "epoch": 0.03415941058271936, "grad_norm": 13.074074844751204, "learning_rate": 4.264214046822743e-07, "loss": 2.8509, "step": 51 }, { "epoch": 0.034829202947086406, "grad_norm": 9.25203175381859, "learning_rate": 4.347826086956522e-07, "loss": 2.6124, "step": 52 }, { "epoch": 0.03549899531145345, "grad_norm": 12.368073919917885, "learning_rate": 4.431438127090301e-07, "loss": 2.9244, "step": 53 }, { "epoch": 0.0361687876758205, "grad_norm": 11.086477716139884, "learning_rate": 4.5150501672240806e-07, "loss": 2.8155, "step": 54 }, { "epoch": 0.036838580040187544, "grad_norm": 15.556605845425212, "learning_rate": 4.5986622073578597e-07, "loss": 2.7958, "step": 55 }, { "epoch": 0.03750837240455459, "grad_norm": 10.238119400380702, "learning_rate": 4.682274247491639e-07, "loss": 2.3461, "step": 56 }, { "epoch": 0.038178164768921635, "grad_norm": 8.08228254031281, "learning_rate": 4.765886287625419e-07, "loss": 2.4209, "step": 57 }, { "epoch": 0.03884795713328868, "grad_norm": 12.84327102014415, "learning_rate": 4.849498327759198e-07, "loss": 3.0494, "step": 58 }, { "epoch": 0.03951774949765573, "grad_norm": 8.961345383061586, "learning_rate": 4.933110367892977e-07, "loss": 2.6697, "step": 59 }, { "epoch": 0.04018754186202277, "grad_norm": 8.713227473274038, "learning_rate": 5.016722408026756e-07, "loss": 2.7243, "step": 60 }, { "epoch": 0.04085733422638982, "grad_norm": 7.486100735606241, "learning_rate": 5.100334448160535e-07, "loss": 2.3621, "step": 61 }, { "epoch": 0.041527126590756865, "grad_norm": 13.74225811514909, "learning_rate": 5.183946488294314e-07, "loss": 2.5102, "step": 62 }, { "epoch": 0.04219691895512391, "grad_norm": 8.417304553964605, "learning_rate": 5.267558528428094e-07, "loss": 2.6936, "step": 63 }, { "epoch": 0.042866711319490956, "grad_norm": 7.624361292947942, "learning_rate": 5.351170568561874e-07, "loss": 2.5782, "step": 64 }, { "epoch": 0.043536503683858, "grad_norm": 8.415987468258116, "learning_rate": 5.434782608695653e-07, "loss": 2.4739, "step": 65 }, { "epoch": 0.04420629604822505, "grad_norm": 6.211926816818595, "learning_rate": 5.518394648829432e-07, "loss": 2.3751, "step": 66 }, { "epoch": 0.044876088412592094, "grad_norm": 6.687278182109561, "learning_rate": 5.602006688963211e-07, "loss": 2.7126, "step": 67 }, { "epoch": 0.04554588077695914, "grad_norm": 11.221201239071942, "learning_rate": 5.68561872909699e-07, "loss": 2.2951, "step": 68 }, { "epoch": 0.046215673141326186, "grad_norm": 6.791267744944243, "learning_rate": 5.76923076923077e-07, "loss": 2.7452, "step": 69 }, { "epoch": 0.04688546550569324, "grad_norm": 7.095535854924889, "learning_rate": 5.852842809364549e-07, "loss": 2.4258, "step": 70 }, { "epoch": 0.047555257870060284, "grad_norm": 6.138386464965501, "learning_rate": 5.936454849498328e-07, "loss": 2.6405, "step": 71 }, { "epoch": 0.04822505023442733, "grad_norm": 5.310785407801751, "learning_rate": 6.020066889632107e-07, "loss": 2.6215, "step": 72 }, { "epoch": 0.048894842598794376, "grad_norm": 5.5867653344526955, "learning_rate": 6.103678929765887e-07, "loss": 2.6725, "step": 73 }, { "epoch": 0.04956463496316142, "grad_norm": 6.2570982195073475, "learning_rate": 6.187290969899666e-07, "loss": 2.3999, "step": 74 }, { "epoch": 0.05023442732752847, "grad_norm": 7.819311478310096, "learning_rate": 6.270903010033446e-07, "loss": 2.5896, "step": 75 }, { "epoch": 0.050904219691895514, "grad_norm": 6.408755811561875, "learning_rate": 6.354515050167225e-07, "loss": 2.5311, "step": 76 }, { "epoch": 0.05157401205626256, "grad_norm": 4.484269211877837, "learning_rate": 6.438127090301004e-07, "loss": 2.3058, "step": 77 }, { "epoch": 0.052243804420629605, "grad_norm": 4.580694814596172, "learning_rate": 6.521739130434783e-07, "loss": 2.4006, "step": 78 }, { "epoch": 0.05291359678499665, "grad_norm": 6.390617591451747, "learning_rate": 6.605351170568562e-07, "loss": 2.6326, "step": 79 }, { "epoch": 0.0535833891493637, "grad_norm": 4.733116881316955, "learning_rate": 6.688963210702341e-07, "loss": 2.5464, "step": 80 }, { "epoch": 0.05425318151373074, "grad_norm": 6.02018802909916, "learning_rate": 6.77257525083612e-07, "loss": 2.5059, "step": 81 }, { "epoch": 0.05492297387809779, "grad_norm": 4.054095061305118, "learning_rate": 6.8561872909699e-07, "loss": 2.1151, "step": 82 }, { "epoch": 0.055592766242464835, "grad_norm": 4.850517497661918, "learning_rate": 6.939799331103679e-07, "loss": 2.5081, "step": 83 }, { "epoch": 0.05626255860683188, "grad_norm": 4.402337106851737, "learning_rate": 7.023411371237459e-07, "loss": 2.3531, "step": 84 }, { "epoch": 0.056932350971198926, "grad_norm": 5.296232257038074, "learning_rate": 7.107023411371238e-07, "loss": 2.5506, "step": 85 }, { "epoch": 0.05760214333556597, "grad_norm": 4.324625364684367, "learning_rate": 7.190635451505017e-07, "loss": 2.4832, "step": 86 }, { "epoch": 0.05827193569993302, "grad_norm": 6.706968077057065, "learning_rate": 7.274247491638796e-07, "loss": 2.4381, "step": 87 }, { "epoch": 0.058941728064300064, "grad_norm": 3.7799413701409925, "learning_rate": 7.357859531772576e-07, "loss": 2.0418, "step": 88 }, { "epoch": 0.05961152042866711, "grad_norm": 4.733398931859532, "learning_rate": 7.441471571906355e-07, "loss": 2.5699, "step": 89 }, { "epoch": 0.06028131279303416, "grad_norm": 4.757660194517577, "learning_rate": 7.525083612040135e-07, "loss": 2.5884, "step": 90 }, { "epoch": 0.06095110515740121, "grad_norm": 5.405550096233397, "learning_rate": 7.608695652173914e-07, "loss": 2.3572, "step": 91 }, { "epoch": 0.061620897521768254, "grad_norm": 4.862518725331412, "learning_rate": 7.692307692307694e-07, "loss": 2.6934, "step": 92 }, { "epoch": 0.0622906898861353, "grad_norm": 4.5028789421370385, "learning_rate": 7.775919732441473e-07, "loss": 2.2655, "step": 93 }, { "epoch": 0.06296048225050234, "grad_norm": 4.563097297917544, "learning_rate": 7.859531772575252e-07, "loss": 2.5964, "step": 94 }, { "epoch": 0.06363027461486939, "grad_norm": 6.402145032768904, "learning_rate": 7.943143812709031e-07, "loss": 2.3076, "step": 95 }, { "epoch": 0.06430006697923643, "grad_norm": 4.499859209356978, "learning_rate": 8.02675585284281e-07, "loss": 2.5495, "step": 96 }, { "epoch": 0.06496985934360348, "grad_norm": 4.143084698311203, "learning_rate": 8.110367892976589e-07, "loss": 2.3349, "step": 97 }, { "epoch": 0.06563965170797052, "grad_norm": 4.35437593150457, "learning_rate": 8.193979933110368e-07, "loss": 2.6001, "step": 98 }, { "epoch": 0.06630944407233758, "grad_norm": 3.890441437064331, "learning_rate": 8.277591973244148e-07, "loss": 2.3291, "step": 99 }, { "epoch": 0.06697923643670463, "grad_norm": 5.802971686195107, "learning_rate": 8.361204013377927e-07, "loss": 2.2001, "step": 100 }, { "epoch": 0.06764902880107167, "grad_norm": 7.206388958682341, "learning_rate": 8.444816053511706e-07, "loss": 2.3388, "step": 101 }, { "epoch": 0.06831882116543872, "grad_norm": 3.868894111209427, "learning_rate": 8.528428093645486e-07, "loss": 2.1697, "step": 102 }, { "epoch": 0.06898861352980576, "grad_norm": 4.522687603612486, "learning_rate": 8.612040133779265e-07, "loss": 2.3985, "step": 103 }, { "epoch": 0.06965840589417281, "grad_norm": 5.016458588380614, "learning_rate": 8.695652173913044e-07, "loss": 2.3063, "step": 104 }, { "epoch": 0.07032819825853985, "grad_norm": 13.687864063808622, "learning_rate": 8.779264214046823e-07, "loss": 2.3258, "step": 105 }, { "epoch": 0.0709979906229069, "grad_norm": 3.950461033358276, "learning_rate": 8.862876254180602e-07, "loss": 2.2051, "step": 106 }, { "epoch": 0.07166778298727394, "grad_norm": 4.428312610534652, "learning_rate": 8.946488294314382e-07, "loss": 2.4124, "step": 107 }, { "epoch": 0.072337575351641, "grad_norm": 4.159304814286087, "learning_rate": 9.030100334448161e-07, "loss": 2.2715, "step": 108 }, { "epoch": 0.07300736771600803, "grad_norm": 4.381950010061395, "learning_rate": 9.11371237458194e-07, "loss": 1.946, "step": 109 }, { "epoch": 0.07367716008037509, "grad_norm": 6.741190836503955, "learning_rate": 9.197324414715719e-07, "loss": 2.3348, "step": 110 }, { "epoch": 0.07434695244474213, "grad_norm": 5.315173478701377, "learning_rate": 9.280936454849498e-07, "loss": 2.1915, "step": 111 }, { "epoch": 0.07501674480910918, "grad_norm": 6.69098823130175, "learning_rate": 9.364548494983278e-07, "loss": 2.2811, "step": 112 }, { "epoch": 0.07568653717347622, "grad_norm": 8.867329811012043, "learning_rate": 9.448160535117059e-07, "loss": 2.2528, "step": 113 }, { "epoch": 0.07635632953784327, "grad_norm": 3.6719627946884694, "learning_rate": 9.531772575250838e-07, "loss": 2.1056, "step": 114 }, { "epoch": 0.07702612190221031, "grad_norm": 5.862951460881078, "learning_rate": 9.615384615384617e-07, "loss": 2.3919, "step": 115 }, { "epoch": 0.07769591426657736, "grad_norm": 4.8365932046479205, "learning_rate": 9.698996655518396e-07, "loss": 2.2752, "step": 116 }, { "epoch": 0.0783657066309444, "grad_norm": 5.741098763124264, "learning_rate": 9.782608695652175e-07, "loss": 2.2493, "step": 117 }, { "epoch": 0.07903549899531145, "grad_norm": 4.5471334993422525, "learning_rate": 9.866220735785954e-07, "loss": 2.2061, "step": 118 }, { "epoch": 0.07970529135967849, "grad_norm": 16.195919803337244, "learning_rate": 9.949832775919733e-07, "loss": 2.3794, "step": 119 }, { "epoch": 0.08037508372404555, "grad_norm": 4.938285271877185, "learning_rate": 1.0033444816053512e-06, "loss": 2.3704, "step": 120 }, { "epoch": 0.0810448760884126, "grad_norm": 4.315484247028453, "learning_rate": 1.0117056856187292e-06, "loss": 2.3134, "step": 121 }, { "epoch": 0.08171466845277964, "grad_norm": 4.182844775230428, "learning_rate": 1.020066889632107e-06, "loss": 2.3008, "step": 122 }, { "epoch": 0.08238446081714669, "grad_norm": 3.894727441237566, "learning_rate": 1.028428093645485e-06, "loss": 2.006, "step": 123 }, { "epoch": 0.08305425318151373, "grad_norm": 4.549843037936009, "learning_rate": 1.0367892976588629e-06, "loss": 2.3595, "step": 124 }, { "epoch": 0.08372404554588078, "grad_norm": 10.157943151723673, "learning_rate": 1.0451505016722408e-06, "loss": 2.5685, "step": 125 }, { "epoch": 0.08439383791024782, "grad_norm": 3.747231113929926, "learning_rate": 1.0535117056856187e-06, "loss": 2.078, "step": 126 }, { "epoch": 0.08506363027461487, "grad_norm": 4.38685764191898, "learning_rate": 1.0618729096989968e-06, "loss": 2.2877, "step": 127 }, { "epoch": 0.08573342263898191, "grad_norm": 3.966527906663706, "learning_rate": 1.0702341137123747e-06, "loss": 2.2565, "step": 128 }, { "epoch": 0.08640321500334897, "grad_norm": 4.756261084792717, "learning_rate": 1.0785953177257526e-06, "loss": 2.1563, "step": 129 }, { "epoch": 0.087073007367716, "grad_norm": 6.0737468740051686, "learning_rate": 1.0869565217391306e-06, "loss": 2.1845, "step": 130 }, { "epoch": 0.08774279973208306, "grad_norm": 9.779919862139783, "learning_rate": 1.0953177257525085e-06, "loss": 2.4369, "step": 131 }, { "epoch": 0.0884125920964501, "grad_norm": 17.06050020098409, "learning_rate": 1.1036789297658864e-06, "loss": 2.2936, "step": 132 }, { "epoch": 0.08908238446081715, "grad_norm": 4.159180502585554, "learning_rate": 1.1120401337792643e-06, "loss": 1.9933, "step": 133 }, { "epoch": 0.08975217682518419, "grad_norm": 6.515768605932017, "learning_rate": 1.1204013377926422e-06, "loss": 2.4507, "step": 134 }, { "epoch": 0.09042196918955124, "grad_norm": 10.248514354027074, "learning_rate": 1.12876254180602e-06, "loss": 2.1969, "step": 135 }, { "epoch": 0.09109176155391828, "grad_norm": 13.330855614969845, "learning_rate": 1.137123745819398e-06, "loss": 2.3026, "step": 136 }, { "epoch": 0.09176155391828533, "grad_norm": 11.789668723991142, "learning_rate": 1.145484949832776e-06, "loss": 2.1759, "step": 137 }, { "epoch": 0.09243134628265237, "grad_norm": 5.655885406591914, "learning_rate": 1.153846153846154e-06, "loss": 2.2147, "step": 138 }, { "epoch": 0.09310113864701942, "grad_norm": 4.044902218073293, "learning_rate": 1.162207357859532e-06, "loss": 2.116, "step": 139 }, { "epoch": 0.09377093101138648, "grad_norm": 4.3990716195176205, "learning_rate": 1.1705685618729099e-06, "loss": 2.1849, "step": 140 }, { "epoch": 0.09444072337575352, "grad_norm": 5.653953890277716, "learning_rate": 1.1789297658862878e-06, "loss": 2.1081, "step": 141 }, { "epoch": 0.09511051574012057, "grad_norm": 4.380281125741834, "learning_rate": 1.1872909698996657e-06, "loss": 2.3069, "step": 142 }, { "epoch": 0.09578030810448761, "grad_norm": 4.417777467045047, "learning_rate": 1.1956521739130436e-06, "loss": 2.3755, "step": 143 }, { "epoch": 0.09645010046885466, "grad_norm": 7.166924975977487, "learning_rate": 1.2040133779264215e-06, "loss": 2.0609, "step": 144 }, { "epoch": 0.0971198928332217, "grad_norm": 4.492024368425302, "learning_rate": 1.2123745819397994e-06, "loss": 2.2098, "step": 145 }, { "epoch": 0.09778968519758875, "grad_norm": 4.521936756726987, "learning_rate": 1.2207357859531773e-06, "loss": 2.1659, "step": 146 }, { "epoch": 0.09845947756195579, "grad_norm": 4.527062681345587, "learning_rate": 1.2290969899665552e-06, "loss": 2.1553, "step": 147 }, { "epoch": 0.09912926992632284, "grad_norm": 8.119330049854637, "learning_rate": 1.2374581939799331e-06, "loss": 2.22, "step": 148 }, { "epoch": 0.09979906229068988, "grad_norm": 6.981855626807168, "learning_rate": 1.245819397993311e-06, "loss": 2.2304, "step": 149 }, { "epoch": 0.10046885465505694, "grad_norm": 4.410240236082328, "learning_rate": 1.2541806020066892e-06, "loss": 2.2929, "step": 150 }, { "epoch": 0.10113864701942397, "grad_norm": 4.654315227921937, "learning_rate": 1.262541806020067e-06, "loss": 2.328, "step": 151 }, { "epoch": 0.10180843938379103, "grad_norm": 5.742461680456988, "learning_rate": 1.270903010033445e-06, "loss": 2.4706, "step": 152 }, { "epoch": 0.10247823174815807, "grad_norm": 4.343977963341409, "learning_rate": 1.2792642140468229e-06, "loss": 2.1292, "step": 153 }, { "epoch": 0.10314802411252512, "grad_norm": 4.374112175620753, "learning_rate": 1.2876254180602008e-06, "loss": 2.27, "step": 154 }, { "epoch": 0.10381781647689216, "grad_norm": 4.227296955589398, "learning_rate": 1.2959866220735787e-06, "loss": 2.2487, "step": 155 }, { "epoch": 0.10448760884125921, "grad_norm": 5.4927189192191355, "learning_rate": 1.3043478260869566e-06, "loss": 2.1705, "step": 156 }, { "epoch": 0.10515740120562625, "grad_norm": 4.448107550318989, "learning_rate": 1.3127090301003345e-06, "loss": 2.3288, "step": 157 }, { "epoch": 0.1058271935699933, "grad_norm": 4.779159822613719, "learning_rate": 1.3210702341137124e-06, "loss": 2.0256, "step": 158 }, { "epoch": 0.10649698593436034, "grad_norm": 4.209002966144807, "learning_rate": 1.3294314381270903e-06, "loss": 2.1058, "step": 159 }, { "epoch": 0.1071667782987274, "grad_norm": 5.6081273389767805, "learning_rate": 1.3377926421404683e-06, "loss": 2.1366, "step": 160 }, { "epoch": 0.10783657066309445, "grad_norm": 4.155107718101385, "learning_rate": 1.3461538461538462e-06, "loss": 2.0868, "step": 161 }, { "epoch": 0.10850636302746149, "grad_norm": 3.9000257642213225, "learning_rate": 1.354515050167224e-06, "loss": 2.0959, "step": 162 }, { "epoch": 0.10917615539182854, "grad_norm": 5.569961107973028, "learning_rate": 1.362876254180602e-06, "loss": 2.3374, "step": 163 }, { "epoch": 0.10984594775619558, "grad_norm": 4.238746017914896, "learning_rate": 1.37123745819398e-06, "loss": 2.1185, "step": 164 }, { "epoch": 0.11051574012056263, "grad_norm": 4.0778371258881245, "learning_rate": 1.3795986622073578e-06, "loss": 2.1994, "step": 165 }, { "epoch": 0.11118553248492967, "grad_norm": 5.735097455717207, "learning_rate": 1.3879598662207357e-06, "loss": 2.0825, "step": 166 }, { "epoch": 0.11185532484929672, "grad_norm": 5.688538673895958, "learning_rate": 1.3963210702341138e-06, "loss": 1.9766, "step": 167 }, { "epoch": 0.11252511721366376, "grad_norm": 4.90490630437972, "learning_rate": 1.4046822742474917e-06, "loss": 2.2804, "step": 168 }, { "epoch": 0.11319490957803081, "grad_norm": 6.320489824764452, "learning_rate": 1.4130434782608697e-06, "loss": 2.2034, "step": 169 }, { "epoch": 0.11386470194239785, "grad_norm": 4.225209727434135, "learning_rate": 1.4214046822742476e-06, "loss": 2.0288, "step": 170 }, { "epoch": 0.1145344943067649, "grad_norm": 4.605078119390726, "learning_rate": 1.4297658862876255e-06, "loss": 1.8097, "step": 171 }, { "epoch": 0.11520428667113194, "grad_norm": 5.1595861961042235, "learning_rate": 1.4381270903010034e-06, "loss": 2.1479, "step": 172 }, { "epoch": 0.115874079035499, "grad_norm": 3.890871145538063, "learning_rate": 1.4464882943143813e-06, "loss": 1.9449, "step": 173 }, { "epoch": 0.11654387139986604, "grad_norm": 4.197063497704128, "learning_rate": 1.4548494983277592e-06, "loss": 1.7902, "step": 174 }, { "epoch": 0.11721366376423309, "grad_norm": 5.684008704194154, "learning_rate": 1.4632107023411373e-06, "loss": 2.0659, "step": 175 }, { "epoch": 0.11788345612860013, "grad_norm": 5.794700023442698, "learning_rate": 1.4715719063545152e-06, "loss": 2.1375, "step": 176 }, { "epoch": 0.11855324849296718, "grad_norm": 5.340123097440851, "learning_rate": 1.4799331103678931e-06, "loss": 2.0713, "step": 177 }, { "epoch": 0.11922304085733422, "grad_norm": 11.469814953950651, "learning_rate": 1.488294314381271e-06, "loss": 2.1541, "step": 178 }, { "epoch": 0.11989283322170127, "grad_norm": 8.099334016026832, "learning_rate": 1.4966555183946492e-06, "loss": 2.362, "step": 179 }, { "epoch": 0.12056262558606833, "grad_norm": 4.246092810397689, "learning_rate": 1.505016722408027e-06, "loss": 2.196, "step": 180 }, { "epoch": 0.12123241795043536, "grad_norm": 4.551840277656565, "learning_rate": 1.513377926421405e-06, "loss": 2.215, "step": 181 }, { "epoch": 0.12190221031480242, "grad_norm": 4.200039717846675, "learning_rate": 1.521739130434783e-06, "loss": 2.0, "step": 182 }, { "epoch": 0.12257200267916946, "grad_norm": 4.939216168103558, "learning_rate": 1.5301003344481608e-06, "loss": 2.1452, "step": 183 }, { "epoch": 0.12324179504353651, "grad_norm": 4.1674706770053325, "learning_rate": 1.5384615384615387e-06, "loss": 2.1771, "step": 184 }, { "epoch": 0.12391158740790355, "grad_norm": 8.976031955670496, "learning_rate": 1.5468227424749166e-06, "loss": 2.0771, "step": 185 }, { "epoch": 0.1245813797722706, "grad_norm": 3.7377280013449288, "learning_rate": 1.5551839464882945e-06, "loss": 1.8825, "step": 186 }, { "epoch": 0.12525117213663764, "grad_norm": 4.406382907491899, "learning_rate": 1.5635451505016724e-06, "loss": 2.0267, "step": 187 }, { "epoch": 0.12592096450100468, "grad_norm": 6.701005106459511, "learning_rate": 1.5719063545150504e-06, "loss": 1.8591, "step": 188 }, { "epoch": 0.12659075686537175, "grad_norm": 3.9839992925775554, "learning_rate": 1.5802675585284283e-06, "loss": 1.8693, "step": 189 }, { "epoch": 0.12726054922973878, "grad_norm": 5.100742538608377, "learning_rate": 1.5886287625418062e-06, "loss": 2.3117, "step": 190 }, { "epoch": 0.12793034159410582, "grad_norm": 4.439890782733566, "learning_rate": 1.596989966555184e-06, "loss": 2.2809, "step": 191 }, { "epoch": 0.12860013395847286, "grad_norm": 5.550065288440506, "learning_rate": 1.605351170568562e-06, "loss": 2.05, "step": 192 }, { "epoch": 0.12926992632283993, "grad_norm": 4.136735928036315, "learning_rate": 1.61371237458194e-06, "loss": 2.1464, "step": 193 }, { "epoch": 0.12993971868720697, "grad_norm": 4.327938483660195, "learning_rate": 1.6220735785953178e-06, "loss": 2.1603, "step": 194 }, { "epoch": 0.130609511051574, "grad_norm": 4.262598058578376, "learning_rate": 1.6304347826086957e-06, "loss": 2.13, "step": 195 }, { "epoch": 0.13127930341594105, "grad_norm": 3.900649755915667, "learning_rate": 1.6387959866220736e-06, "loss": 2.0204, "step": 196 }, { "epoch": 0.1319490957803081, "grad_norm": 4.493594548505599, "learning_rate": 1.6471571906354518e-06, "loss": 1.8568, "step": 197 }, { "epoch": 0.13261888814467515, "grad_norm": 4.423183806210867, "learning_rate": 1.6555183946488297e-06, "loss": 2.0326, "step": 198 }, { "epoch": 0.1332886805090422, "grad_norm": 5.428227342362131, "learning_rate": 1.6638795986622076e-06, "loss": 2.0324, "step": 199 }, { "epoch": 0.13395847287340926, "grad_norm": 3.7804281856625046, "learning_rate": 1.6722408026755855e-06, "loss": 2.022, "step": 200 }, { "epoch": 0.1346282652377763, "grad_norm": 4.429304957740435, "learning_rate": 1.6806020066889634e-06, "loss": 2.1632, "step": 201 }, { "epoch": 0.13529805760214333, "grad_norm": 24.84393845684732, "learning_rate": 1.6889632107023413e-06, "loss": 1.9586, "step": 202 }, { "epoch": 0.13596784996651037, "grad_norm": 5.355687503356123, "learning_rate": 1.6973244147157192e-06, "loss": 2.1212, "step": 203 }, { "epoch": 0.13663764233087744, "grad_norm": 5.622054795549618, "learning_rate": 1.7056856187290971e-06, "loss": 1.9181, "step": 204 }, { "epoch": 0.13730743469524448, "grad_norm": 4.22675895089208, "learning_rate": 1.714046822742475e-06, "loss": 2.2174, "step": 205 }, { "epoch": 0.13797722705961152, "grad_norm": 3.9556423364180158, "learning_rate": 1.722408026755853e-06, "loss": 2.1604, "step": 206 }, { "epoch": 0.13864701942397856, "grad_norm": 9.215368475776009, "learning_rate": 1.7307692307692308e-06, "loss": 1.9173, "step": 207 }, { "epoch": 0.13931681178834562, "grad_norm": 3.8563993928098155, "learning_rate": 1.7391304347826088e-06, "loss": 2.1638, "step": 208 }, { "epoch": 0.13998660415271266, "grad_norm": 3.6951017294409017, "learning_rate": 1.7474916387959867e-06, "loss": 1.8314, "step": 209 }, { "epoch": 0.1406563965170797, "grad_norm": 4.016890997688114, "learning_rate": 1.7558528428093646e-06, "loss": 2.0568, "step": 210 }, { "epoch": 0.14132618888144674, "grad_norm": 3.9756483098172266, "learning_rate": 1.7642140468227425e-06, "loss": 2.0611, "step": 211 }, { "epoch": 0.1419959812458138, "grad_norm": 5.322063846305247, "learning_rate": 1.7725752508361204e-06, "loss": 2.0243, "step": 212 }, { "epoch": 0.14266577361018085, "grad_norm": 3.494819241910969, "learning_rate": 1.7809364548494983e-06, "loss": 2.0094, "step": 213 }, { "epoch": 0.14333556597454788, "grad_norm": 4.166040405370835, "learning_rate": 1.7892976588628764e-06, "loss": 2.1127, "step": 214 }, { "epoch": 0.14400535833891492, "grad_norm": 6.176207632381695, "learning_rate": 1.7976588628762543e-06, "loss": 2.1676, "step": 215 }, { "epoch": 0.144675150703282, "grad_norm": 3.7384931525802587, "learning_rate": 1.8060200668896322e-06, "loss": 2.1949, "step": 216 }, { "epoch": 0.14534494306764903, "grad_norm": 3.448755373799391, "learning_rate": 1.8143812709030102e-06, "loss": 2.1368, "step": 217 }, { "epoch": 0.14601473543201607, "grad_norm": 3.840771228185559, "learning_rate": 1.822742474916388e-06, "loss": 2.1888, "step": 218 }, { "epoch": 0.14668452779638314, "grad_norm": 4.377472651291213, "learning_rate": 1.831103678929766e-06, "loss": 2.2062, "step": 219 }, { "epoch": 0.14735432016075017, "grad_norm": 3.5241500019927847, "learning_rate": 1.8394648829431439e-06, "loss": 1.9209, "step": 220 }, { "epoch": 0.1480241125251172, "grad_norm": 3.42848303682146, "learning_rate": 1.8478260869565218e-06, "loss": 2.0964, "step": 221 }, { "epoch": 0.14869390488948425, "grad_norm": 3.2989033017608715, "learning_rate": 1.8561872909698997e-06, "loss": 2.0062, "step": 222 }, { "epoch": 0.14936369725385132, "grad_norm": 9.905320058265135, "learning_rate": 1.8645484949832776e-06, "loss": 1.9428, "step": 223 }, { "epoch": 0.15003348961821836, "grad_norm": 3.7561505956473904, "learning_rate": 1.8729096989966555e-06, "loss": 2.0865, "step": 224 }, { "epoch": 0.1507032819825854, "grad_norm": 3.275155977029667, "learning_rate": 1.8812709030100336e-06, "loss": 2.1356, "step": 225 }, { "epoch": 0.15137307434695244, "grad_norm": 3.0321773072536407, "learning_rate": 1.8896321070234118e-06, "loss": 1.7617, "step": 226 }, { "epoch": 0.1520428667113195, "grad_norm": 5.16189463068424, "learning_rate": 1.8979933110367897e-06, "loss": 1.7266, "step": 227 }, { "epoch": 0.15271265907568654, "grad_norm": 3.5992942017943794, "learning_rate": 1.9063545150501676e-06, "loss": 1.9981, "step": 228 }, { "epoch": 0.15338245144005358, "grad_norm": 5.671149738281153, "learning_rate": 1.9147157190635453e-06, "loss": 1.6819, "step": 229 }, { "epoch": 0.15405224380442062, "grad_norm": 3.310610549208023, "learning_rate": 1.9230769230769234e-06, "loss": 2.0287, "step": 230 }, { "epoch": 0.15472203616878769, "grad_norm": 3.0904840576150954, "learning_rate": 1.931438127090301e-06, "loss": 1.7779, "step": 231 }, { "epoch": 0.15539182853315472, "grad_norm": 3.112591456778929, "learning_rate": 1.9397993311036792e-06, "loss": 1.8921, "step": 232 }, { "epoch": 0.15606162089752176, "grad_norm": 4.561960615950001, "learning_rate": 1.948160535117057e-06, "loss": 2.0174, "step": 233 }, { "epoch": 0.1567314132618888, "grad_norm": 3.7345200620351515, "learning_rate": 1.956521739130435e-06, "loss": 1.7625, "step": 234 }, { "epoch": 0.15740120562625587, "grad_norm": 3.012711674125082, "learning_rate": 1.964882943143813e-06, "loss": 1.927, "step": 235 }, { "epoch": 0.1580709979906229, "grad_norm": 3.997423142312092, "learning_rate": 1.973244147157191e-06, "loss": 2.0197, "step": 236 }, { "epoch": 0.15874079035498995, "grad_norm": 7.3262786372681274, "learning_rate": 1.981605351170569e-06, "loss": 1.9615, "step": 237 }, { "epoch": 0.15941058271935699, "grad_norm": 3.0859182411567674, "learning_rate": 1.9899665551839467e-06, "loss": 2.0162, "step": 238 }, { "epoch": 0.16008037508372405, "grad_norm": 3.193168557204858, "learning_rate": 1.998327759197325e-06, "loss": 1.9329, "step": 239 }, { "epoch": 0.1607501674480911, "grad_norm": 4.673795125783191, "learning_rate": 2.0066889632107025e-06, "loss": 1.9542, "step": 240 }, { "epoch": 0.16141995981245813, "grad_norm": 6.309957958387706, "learning_rate": 2.0150501672240806e-06, "loss": 2.0351, "step": 241 }, { "epoch": 0.1620897521768252, "grad_norm": 9.29499414190798, "learning_rate": 2.0234113712374583e-06, "loss": 1.9387, "step": 242 }, { "epoch": 0.16275954454119224, "grad_norm": 3.38911484987636, "learning_rate": 2.0317725752508364e-06, "loss": 2.077, "step": 243 }, { "epoch": 0.16342933690555927, "grad_norm": 10.014097503531366, "learning_rate": 2.040133779264214e-06, "loss": 1.9622, "step": 244 }, { "epoch": 0.1640991292699263, "grad_norm": 3.202693274441352, "learning_rate": 2.0484949832775922e-06, "loss": 1.5992, "step": 245 }, { "epoch": 0.16476892163429338, "grad_norm": 3.391891650567158, "learning_rate": 2.05685618729097e-06, "loss": 1.8131, "step": 246 }, { "epoch": 0.16543871399866042, "grad_norm": 6.122228237398313, "learning_rate": 2.065217391304348e-06, "loss": 1.7337, "step": 247 }, { "epoch": 0.16610850636302746, "grad_norm": 4.051358973928674, "learning_rate": 2.0735785953177258e-06, "loss": 2.0006, "step": 248 }, { "epoch": 0.1667782987273945, "grad_norm": 3.0380702214568927, "learning_rate": 2.081939799331104e-06, "loss": 2.0583, "step": 249 }, { "epoch": 0.16744809109176156, "grad_norm": 3.1851712606544704, "learning_rate": 2.0903010033444816e-06, "loss": 1.9248, "step": 250 }, { "epoch": 0.1681178834561286, "grad_norm": 4.709450958922658, "learning_rate": 2.0986622073578597e-06, "loss": 1.9859, "step": 251 }, { "epoch": 0.16878767582049564, "grad_norm": 2.998430024772106, "learning_rate": 2.1070234113712374e-06, "loss": 1.9375, "step": 252 }, { "epoch": 0.16945746818486268, "grad_norm": 6.37327546684449, "learning_rate": 2.1153846153846155e-06, "loss": 2.0356, "step": 253 }, { "epoch": 0.17012726054922975, "grad_norm": 5.900180040008986, "learning_rate": 2.1237458193979936e-06, "loss": 1.7775, "step": 254 }, { "epoch": 0.1707970529135968, "grad_norm": 2.6691518714770908, "learning_rate": 2.1321070234113713e-06, "loss": 1.8235, "step": 255 }, { "epoch": 0.17146684527796383, "grad_norm": 2.9425744167533385, "learning_rate": 2.1404682274247495e-06, "loss": 1.9991, "step": 256 }, { "epoch": 0.17213663764233086, "grad_norm": 3.1275398947931734, "learning_rate": 2.148829431438127e-06, "loss": 1.8532, "step": 257 }, { "epoch": 0.17280643000669793, "grad_norm": 3.0571494102092682, "learning_rate": 2.1571906354515053e-06, "loss": 2.0989, "step": 258 }, { "epoch": 0.17347622237106497, "grad_norm": 2.614198629391264, "learning_rate": 2.165551839464883e-06, "loss": 1.8527, "step": 259 }, { "epoch": 0.174146014735432, "grad_norm": 5.496948746904526, "learning_rate": 2.173913043478261e-06, "loss": 1.8343, "step": 260 }, { "epoch": 0.17481580709979908, "grad_norm": 5.306606943679884, "learning_rate": 2.182274247491639e-06, "loss": 1.7811, "step": 261 }, { "epoch": 0.17548559946416611, "grad_norm": 4.144235395114425, "learning_rate": 2.190635451505017e-06, "loss": 2.07, "step": 262 }, { "epoch": 0.17615539182853315, "grad_norm": 4.494958599528628, "learning_rate": 2.1989966555183946e-06, "loss": 1.8997, "step": 263 }, { "epoch": 0.1768251841929002, "grad_norm": 3.9496744468886047, "learning_rate": 2.2073578595317727e-06, "loss": 2.0033, "step": 264 }, { "epoch": 0.17749497655726726, "grad_norm": 2.8421909238689524, "learning_rate": 2.2157190635451504e-06, "loss": 1.9149, "step": 265 }, { "epoch": 0.1781647689216343, "grad_norm": 2.761259033482306, "learning_rate": 2.2240802675585286e-06, "loss": 1.9381, "step": 266 }, { "epoch": 0.17883456128600134, "grad_norm": 2.7274593849242312, "learning_rate": 2.2324414715719063e-06, "loss": 2.0722, "step": 267 }, { "epoch": 0.17950435365036838, "grad_norm": 3.0569466623856743, "learning_rate": 2.2408026755852844e-06, "loss": 1.6922, "step": 268 }, { "epoch": 0.18017414601473544, "grad_norm": 4.044950829337992, "learning_rate": 2.249163879598662e-06, "loss": 1.9363, "step": 269 }, { "epoch": 0.18084393837910248, "grad_norm": 2.6465073321015855, "learning_rate": 2.25752508361204e-06, "loss": 2.107, "step": 270 }, { "epoch": 0.18151373074346952, "grad_norm": 6.464300291198303, "learning_rate": 2.2658862876254183e-06, "loss": 1.8636, "step": 271 }, { "epoch": 0.18218352310783656, "grad_norm": 2.7762055718037475, "learning_rate": 2.274247491638796e-06, "loss": 2.0445, "step": 272 }, { "epoch": 0.18285331547220363, "grad_norm": 4.272216221256872, "learning_rate": 2.282608695652174e-06, "loss": 1.7061, "step": 273 }, { "epoch": 0.18352310783657066, "grad_norm": 2.671372727373494, "learning_rate": 2.290969899665552e-06, "loss": 1.9796, "step": 274 }, { "epoch": 0.1841929002009377, "grad_norm": 2.5693669962753893, "learning_rate": 2.29933110367893e-06, "loss": 2.1146, "step": 275 }, { "epoch": 0.18486269256530474, "grad_norm": 3.3922512078896805, "learning_rate": 2.307692307692308e-06, "loss": 1.902, "step": 276 }, { "epoch": 0.1855324849296718, "grad_norm": 4.479685433712042, "learning_rate": 2.3160535117056858e-06, "loss": 1.7188, "step": 277 }, { "epoch": 0.18620227729403885, "grad_norm": 2.621420829859747, "learning_rate": 2.324414715719064e-06, "loss": 2.1364, "step": 278 }, { "epoch": 0.1868720696584059, "grad_norm": 2.433589509374516, "learning_rate": 2.3327759197324416e-06, "loss": 1.8479, "step": 279 }, { "epoch": 0.18754186202277295, "grad_norm": 8.654065891719505, "learning_rate": 2.3411371237458197e-06, "loss": 1.6533, "step": 280 }, { "epoch": 0.18821165438714, "grad_norm": 5.0876535696812955, "learning_rate": 2.3494983277591974e-06, "loss": 1.8605, "step": 281 }, { "epoch": 0.18888144675150703, "grad_norm": 3.1354435282304034, "learning_rate": 2.3578595317725755e-06, "loss": 1.7217, "step": 282 }, { "epoch": 0.18955123911587407, "grad_norm": 2.453434874245602, "learning_rate": 2.3662207357859537e-06, "loss": 1.8706, "step": 283 }, { "epoch": 0.19022103148024114, "grad_norm": 3.3056998576953895, "learning_rate": 2.3745819397993314e-06, "loss": 1.8427, "step": 284 }, { "epoch": 0.19089082384460818, "grad_norm": 2.761565378124557, "learning_rate": 2.3829431438127095e-06, "loss": 1.995, "step": 285 }, { "epoch": 0.19156061620897522, "grad_norm": 3.8503177252214305, "learning_rate": 2.391304347826087e-06, "loss": 2.0989, "step": 286 }, { "epoch": 0.19223040857334225, "grad_norm": 3.414621628888019, "learning_rate": 2.3996655518394653e-06, "loss": 1.7624, "step": 287 }, { "epoch": 0.19290020093770932, "grad_norm": 4.186641436709834, "learning_rate": 2.408026755852843e-06, "loss": 1.9499, "step": 288 }, { "epoch": 0.19356999330207636, "grad_norm": 4.151600873300877, "learning_rate": 2.416387959866221e-06, "loss": 1.9447, "step": 289 }, { "epoch": 0.1942397856664434, "grad_norm": 2.6290425462815223, "learning_rate": 2.424749163879599e-06, "loss": 1.79, "step": 290 }, { "epoch": 0.19490957803081044, "grad_norm": 4.9410446491221, "learning_rate": 2.433110367892977e-06, "loss": 1.9593, "step": 291 }, { "epoch": 0.1955793703951775, "grad_norm": 2.611439508085505, "learning_rate": 2.4414715719063546e-06, "loss": 2.0176, "step": 292 }, { "epoch": 0.19624916275954454, "grad_norm": 4.538059749550721, "learning_rate": 2.4498327759197327e-06, "loss": 1.9432, "step": 293 }, { "epoch": 0.19691895512391158, "grad_norm": 4.994955519112718, "learning_rate": 2.4581939799331104e-06, "loss": 2.0093, "step": 294 }, { "epoch": 0.19758874748827862, "grad_norm": 6.755025426738264, "learning_rate": 2.4665551839464886e-06, "loss": 1.8621, "step": 295 }, { "epoch": 0.1982585398526457, "grad_norm": 2.4828877911033875, "learning_rate": 2.4749163879598663e-06, "loss": 1.8332, "step": 296 }, { "epoch": 0.19892833221701273, "grad_norm": 3.8770261438105034, "learning_rate": 2.4832775919732444e-06, "loss": 1.8694, "step": 297 }, { "epoch": 0.19959812458137977, "grad_norm": 3.362838662235264, "learning_rate": 2.491638795986622e-06, "loss": 1.8709, "step": 298 }, { "epoch": 0.20026791694574683, "grad_norm": 2.6663590514491644, "learning_rate": 2.5e-06, "loss": 1.6635, "step": 299 }, { "epoch": 0.20093770931011387, "grad_norm": 5.560974207871311, "learning_rate": 2.5083612040133783e-06, "loss": 1.9167, "step": 300 }, { "epoch": 0.2016075016744809, "grad_norm": 3.39094807049767, "learning_rate": 2.516722408026756e-06, "loss": 1.7757, "step": 301 }, { "epoch": 0.20227729403884795, "grad_norm": 2.493423718564741, "learning_rate": 2.525083612040134e-06, "loss": 2.0475, "step": 302 }, { "epoch": 0.20294708640321502, "grad_norm": 2.5315197627618784, "learning_rate": 2.533444816053512e-06, "loss": 2.1702, "step": 303 }, { "epoch": 0.20361687876758205, "grad_norm": 4.3077786550841, "learning_rate": 2.54180602006689e-06, "loss": 1.7199, "step": 304 }, { "epoch": 0.2042866711319491, "grad_norm": 2.5846700129774325, "learning_rate": 2.5501672240802677e-06, "loss": 1.8587, "step": 305 }, { "epoch": 0.20495646349631613, "grad_norm": 2.5575771414920614, "learning_rate": 2.5585284280936458e-06, "loss": 1.8177, "step": 306 }, { "epoch": 0.2056262558606832, "grad_norm": 2.4483820209366502, "learning_rate": 2.5668896321070235e-06, "loss": 1.9801, "step": 307 }, { "epoch": 0.20629604822505024, "grad_norm": 4.196730482370021, "learning_rate": 2.5752508361204016e-06, "loss": 1.9101, "step": 308 }, { "epoch": 0.20696584058941728, "grad_norm": 3.0419005829153187, "learning_rate": 2.5836120401337793e-06, "loss": 1.792, "step": 309 }, { "epoch": 0.20763563295378432, "grad_norm": 2.48489021075996, "learning_rate": 2.5919732441471574e-06, "loss": 1.9537, "step": 310 }, { "epoch": 0.20830542531815138, "grad_norm": 2.4925685478347996, "learning_rate": 2.600334448160535e-06, "loss": 2.0294, "step": 311 }, { "epoch": 0.20897521768251842, "grad_norm": 2.414312958337348, "learning_rate": 2.6086956521739132e-06, "loss": 1.7538, "step": 312 }, { "epoch": 0.20964501004688546, "grad_norm": 2.385612398905679, "learning_rate": 2.617056856187291e-06, "loss": 2.0132, "step": 313 }, { "epoch": 0.2103148024112525, "grad_norm": 2.5541573465022345, "learning_rate": 2.625418060200669e-06, "loss": 1.8855, "step": 314 }, { "epoch": 0.21098459477561957, "grad_norm": 2.5455950870887736, "learning_rate": 2.6337792642140468e-06, "loss": 1.9494, "step": 315 }, { "epoch": 0.2116543871399866, "grad_norm": 2.412467501764056, "learning_rate": 2.642140468227425e-06, "loss": 2.0642, "step": 316 }, { "epoch": 0.21232417950435364, "grad_norm": 3.637015159299892, "learning_rate": 2.650501672240803e-06, "loss": 1.6933, "step": 317 }, { "epoch": 0.21299397186872068, "grad_norm": 2.6769507666312373, "learning_rate": 2.6588628762541807e-06, "loss": 1.9756, "step": 318 }, { "epoch": 0.21366376423308775, "grad_norm": 3.0121936005823766, "learning_rate": 2.667224080267559e-06, "loss": 1.9153, "step": 319 }, { "epoch": 0.2143335565974548, "grad_norm": 2.514507775145898, "learning_rate": 2.6755852842809365e-06, "loss": 2.0051, "step": 320 }, { "epoch": 0.21500334896182183, "grad_norm": 2.5394877076478055, "learning_rate": 2.6839464882943146e-06, "loss": 1.8057, "step": 321 }, { "epoch": 0.2156731413261889, "grad_norm": 2.5160191904850278, "learning_rate": 2.6923076923076923e-06, "loss": 1.9716, "step": 322 }, { "epoch": 0.21634293369055593, "grad_norm": 2.5656209788799895, "learning_rate": 2.7006688963210705e-06, "loss": 1.901, "step": 323 }, { "epoch": 0.21701272605492297, "grad_norm": 2.6473104347190928, "learning_rate": 2.709030100334448e-06, "loss": 1.8473, "step": 324 }, { "epoch": 0.21768251841929, "grad_norm": 2.4360678288681843, "learning_rate": 2.7173913043478263e-06, "loss": 1.9774, "step": 325 }, { "epoch": 0.21835231078365708, "grad_norm": 7.161294001817938, "learning_rate": 2.725752508361204e-06, "loss": 1.9964, "step": 326 }, { "epoch": 0.21902210314802412, "grad_norm": 4.2339957718591, "learning_rate": 2.734113712374582e-06, "loss": 1.7606, "step": 327 }, { "epoch": 0.21969189551239116, "grad_norm": 3.6387873681783387, "learning_rate": 2.74247491638796e-06, "loss": 2.0423, "step": 328 }, { "epoch": 0.2203616878767582, "grad_norm": 2.5041997692954863, "learning_rate": 2.750836120401338e-06, "loss": 1.9112, "step": 329 }, { "epoch": 0.22103148024112526, "grad_norm": 2.680178323144739, "learning_rate": 2.7591973244147156e-06, "loss": 2.1118, "step": 330 }, { "epoch": 0.2217012726054923, "grad_norm": 2.3324971152368916, "learning_rate": 2.7675585284280937e-06, "loss": 2.0708, "step": 331 }, { "epoch": 0.22237106496985934, "grad_norm": 6.11001434968708, "learning_rate": 2.7759197324414714e-06, "loss": 1.792, "step": 332 }, { "epoch": 0.22304085733422638, "grad_norm": 3.997497459668957, "learning_rate": 2.7842809364548495e-06, "loss": 1.9926, "step": 333 }, { "epoch": 0.22371064969859344, "grad_norm": 3.82283509761006, "learning_rate": 2.7926421404682277e-06, "loss": 1.8484, "step": 334 }, { "epoch": 0.22438044206296048, "grad_norm": 3.8978113217968735, "learning_rate": 2.8010033444816054e-06, "loss": 1.7986, "step": 335 }, { "epoch": 0.22505023442732752, "grad_norm": 3.5979325364286407, "learning_rate": 2.8093645484949835e-06, "loss": 1.9062, "step": 336 }, { "epoch": 0.22572002679169456, "grad_norm": 3.943872829339072, "learning_rate": 2.817725752508361e-06, "loss": 1.7928, "step": 337 }, { "epoch": 0.22638981915606163, "grad_norm": 3.1948347387749867, "learning_rate": 2.8260869565217393e-06, "loss": 1.7486, "step": 338 }, { "epoch": 0.22705961152042867, "grad_norm": 4.010707046293059, "learning_rate": 2.834448160535117e-06, "loss": 1.8568, "step": 339 }, { "epoch": 0.2277294038847957, "grad_norm": 2.521146306225321, "learning_rate": 2.842809364548495e-06, "loss": 1.8391, "step": 340 }, { "epoch": 0.22839919624916277, "grad_norm": 3.3539733750178717, "learning_rate": 2.851170568561873e-06, "loss": 1.9404, "step": 341 }, { "epoch": 0.2290689886135298, "grad_norm": 2.5601532386691668, "learning_rate": 2.859531772575251e-06, "loss": 1.9472, "step": 342 }, { "epoch": 0.22973878097789685, "grad_norm": 2.662918811192414, "learning_rate": 2.8678929765886286e-06, "loss": 2.1149, "step": 343 }, { "epoch": 0.2304085733422639, "grad_norm": 4.482338690919094, "learning_rate": 2.8762541806020068e-06, "loss": 1.6746, "step": 344 }, { "epoch": 0.23107836570663096, "grad_norm": 2.4615547755955896, "learning_rate": 2.8846153846153845e-06, "loss": 2.1212, "step": 345 }, { "epoch": 0.231748158070998, "grad_norm": 2.9926399645458015, "learning_rate": 2.8929765886287626e-06, "loss": 1.556, "step": 346 }, { "epoch": 0.23241795043536503, "grad_norm": 2.5364319405106874, "learning_rate": 2.9013377926421403e-06, "loss": 1.7487, "step": 347 }, { "epoch": 0.23308774279973207, "grad_norm": 3.713568488653115, "learning_rate": 2.9096989966555184e-06, "loss": 1.8359, "step": 348 }, { "epoch": 0.23375753516409914, "grad_norm": 2.2900487729680505, "learning_rate": 2.918060200668897e-06, "loss": 1.8661, "step": 349 }, { "epoch": 0.23442732752846618, "grad_norm": 2.8117354396882406, "learning_rate": 2.9264214046822746e-06, "loss": 2.0218, "step": 350 }, { "epoch": 0.23509711989283322, "grad_norm": 3.9290725918805602, "learning_rate": 2.9347826086956528e-06, "loss": 1.8982, "step": 351 }, { "epoch": 0.23576691225720026, "grad_norm": 2.3983696682951248, "learning_rate": 2.9431438127090305e-06, "loss": 1.8937, "step": 352 }, { "epoch": 0.23643670462156732, "grad_norm": 2.287215866152363, "learning_rate": 2.9515050167224086e-06, "loss": 1.805, "step": 353 }, { "epoch": 0.23710649698593436, "grad_norm": 2.4167922634262777, "learning_rate": 2.9598662207357863e-06, "loss": 1.8973, "step": 354 }, { "epoch": 0.2377762893503014, "grad_norm": 2.960920181678026, "learning_rate": 2.9682274247491644e-06, "loss": 1.897, "step": 355 }, { "epoch": 0.23844608171466844, "grad_norm": 3.2261953884821217, "learning_rate": 2.976588628762542e-06, "loss": 1.8492, "step": 356 }, { "epoch": 0.2391158740790355, "grad_norm": 4.682025735457098, "learning_rate": 2.9849498327759202e-06, "loss": 1.9073, "step": 357 }, { "epoch": 0.23978566644340255, "grad_norm": 4.475939909595335, "learning_rate": 2.9933110367892983e-06, "loss": 1.8908, "step": 358 }, { "epoch": 0.24045545880776958, "grad_norm": 5.527364216718183, "learning_rate": 3.001672240802676e-06, "loss": 2.2523, "step": 359 }, { "epoch": 0.24112525117213665, "grad_norm": 2.971168745524515, "learning_rate": 3.010033444816054e-06, "loss": 1.7673, "step": 360 }, { "epoch": 0.2417950435365037, "grad_norm": 2.785536987691202, "learning_rate": 3.018394648829432e-06, "loss": 1.6808, "step": 361 }, { "epoch": 0.24246483590087073, "grad_norm": 6.0918115801819495, "learning_rate": 3.02675585284281e-06, "loss": 1.667, "step": 362 }, { "epoch": 0.24313462826523777, "grad_norm": 3.908791996791965, "learning_rate": 3.0351170568561877e-06, "loss": 2.0434, "step": 363 }, { "epoch": 0.24380442062960483, "grad_norm": 3.7098841701040737, "learning_rate": 3.043478260869566e-06, "loss": 1.7684, "step": 364 }, { "epoch": 0.24447421299397187, "grad_norm": 4.600630903978546, "learning_rate": 3.0518394648829435e-06, "loss": 1.8004, "step": 365 }, { "epoch": 0.2451440053583389, "grad_norm": 3.076121989210127, "learning_rate": 3.0602006688963216e-06, "loss": 1.8865, "step": 366 }, { "epoch": 0.24581379772270595, "grad_norm": 8.140817899596255, "learning_rate": 3.0685618729096993e-06, "loss": 1.6498, "step": 367 }, { "epoch": 0.24648359008707302, "grad_norm": 3.6254104605461626, "learning_rate": 3.0769230769230774e-06, "loss": 1.9604, "step": 368 }, { "epoch": 0.24715338245144006, "grad_norm": 3.153088669576466, "learning_rate": 3.085284280936455e-06, "loss": 1.7503, "step": 369 }, { "epoch": 0.2478231748158071, "grad_norm": 2.488871104863748, "learning_rate": 3.0936454849498333e-06, "loss": 1.9337, "step": 370 }, { "epoch": 0.24849296718017413, "grad_norm": 2.2528288064360233, "learning_rate": 3.102006688963211e-06, "loss": 2.0494, "step": 371 }, { "epoch": 0.2491627595445412, "grad_norm": 4.630480301389395, "learning_rate": 3.110367892976589e-06, "loss": 1.7731, "step": 372 }, { "epoch": 0.24983255190890824, "grad_norm": 6.131628033739213, "learning_rate": 3.1187290969899668e-06, "loss": 1.8195, "step": 373 }, { "epoch": 0.2505023442732753, "grad_norm": 4.8042442702742605, "learning_rate": 3.127090301003345e-06, "loss": 1.9552, "step": 374 }, { "epoch": 0.2511721366376423, "grad_norm": 5.445939594505421, "learning_rate": 3.1354515050167226e-06, "loss": 1.4168, "step": 375 }, { "epoch": 0.25184192900200936, "grad_norm": 2.443706048309502, "learning_rate": 3.1438127090301007e-06, "loss": 1.816, "step": 376 }, { "epoch": 0.2525117213663764, "grad_norm": 2.4128267039657465, "learning_rate": 3.152173913043479e-06, "loss": 2.0465, "step": 377 }, { "epoch": 0.2531815137307435, "grad_norm": 4.908305264838648, "learning_rate": 3.1605351170568565e-06, "loss": 1.6888, "step": 378 }, { "epoch": 0.25385130609511053, "grad_norm": 6.8733930078412575, "learning_rate": 3.1688963210702347e-06, "loss": 1.8305, "step": 379 }, { "epoch": 0.25452109845947757, "grad_norm": 4.175859896466855, "learning_rate": 3.1772575250836123e-06, "loss": 1.91, "step": 380 }, { "epoch": 0.2551908908238446, "grad_norm": 3.531787262470253, "learning_rate": 3.1856187290969905e-06, "loss": 1.9012, "step": 381 }, { "epoch": 0.25586068318821165, "grad_norm": 2.6250902956532576, "learning_rate": 3.193979933110368e-06, "loss": 1.7184, "step": 382 }, { "epoch": 0.2565304755525787, "grad_norm": 2.266878569101977, "learning_rate": 3.2023411371237463e-06, "loss": 1.9213, "step": 383 }, { "epoch": 0.2572002679169457, "grad_norm": 2.42697811981765, "learning_rate": 3.210702341137124e-06, "loss": 1.761, "step": 384 }, { "epoch": 0.2578700602813128, "grad_norm": 2.1342672570807117, "learning_rate": 3.219063545150502e-06, "loss": 1.7818, "step": 385 }, { "epoch": 0.25853985264567986, "grad_norm": 2.3122529881852327, "learning_rate": 3.22742474916388e-06, "loss": 1.769, "step": 386 }, { "epoch": 0.2592096450100469, "grad_norm": 2.4203408317003623, "learning_rate": 3.235785953177258e-06, "loss": 2.0264, "step": 387 }, { "epoch": 0.25987943737441394, "grad_norm": 2.437653375106628, "learning_rate": 3.2441471571906356e-06, "loss": 1.8327, "step": 388 }, { "epoch": 0.260549229738781, "grad_norm": 3.0829651382320455, "learning_rate": 3.2525083612040137e-06, "loss": 1.7828, "step": 389 }, { "epoch": 0.261219022103148, "grad_norm": 3.530364191880072, "learning_rate": 3.2608695652173914e-06, "loss": 1.6801, "step": 390 }, { "epoch": 0.26188881446751505, "grad_norm": 2.30406621956897, "learning_rate": 3.2692307692307696e-06, "loss": 1.8209, "step": 391 }, { "epoch": 0.2625586068318821, "grad_norm": 2.9391205590139955, "learning_rate": 3.2775919732441473e-06, "loss": 2.0861, "step": 392 }, { "epoch": 0.2632283991962492, "grad_norm": 3.9523407796083196, "learning_rate": 3.2859531772575254e-06, "loss": 1.6146, "step": 393 }, { "epoch": 0.2638981915606162, "grad_norm": 3.3164110481971307, "learning_rate": 3.2943143812709035e-06, "loss": 1.9044, "step": 394 }, { "epoch": 0.26456798392498326, "grad_norm": 2.3518948599764675, "learning_rate": 3.302675585284281e-06, "loss": 1.6844, "step": 395 }, { "epoch": 0.2652377762893503, "grad_norm": 2.9549720166568525, "learning_rate": 3.3110367892976593e-06, "loss": 1.7153, "step": 396 }, { "epoch": 0.26590756865371734, "grad_norm": 3.375052659801266, "learning_rate": 3.319397993311037e-06, "loss": 1.9298, "step": 397 }, { "epoch": 0.2665773610180844, "grad_norm": 5.7269890312325495, "learning_rate": 3.327759197324415e-06, "loss": 1.8653, "step": 398 }, { "epoch": 0.2672471533824514, "grad_norm": 2.994410977392051, "learning_rate": 3.336120401337793e-06, "loss": 1.9254, "step": 399 }, { "epoch": 0.2679169457468185, "grad_norm": 2.4423211003612546, "learning_rate": 3.344481605351171e-06, "loss": 1.995, "step": 400 }, { "epoch": 0.26858673811118555, "grad_norm": 2.3533504431959416, "learning_rate": 3.3528428093645487e-06, "loss": 1.8441, "step": 401 }, { "epoch": 0.2692565304755526, "grad_norm": 3.2224388916682316, "learning_rate": 3.3612040133779268e-06, "loss": 1.9118, "step": 402 }, { "epoch": 0.26992632283991963, "grad_norm": 2.4643984232777436, "learning_rate": 3.3695652173913045e-06, "loss": 1.6822, "step": 403 }, { "epoch": 0.27059611520428667, "grad_norm": 5.342430461357322, "learning_rate": 3.3779264214046826e-06, "loss": 1.9098, "step": 404 }, { "epoch": 0.2712659075686537, "grad_norm": 5.460416823868915, "learning_rate": 3.3862876254180603e-06, "loss": 1.8865, "step": 405 }, { "epoch": 0.27193569993302075, "grad_norm": 3.5982528843585544, "learning_rate": 3.3946488294314384e-06, "loss": 1.8738, "step": 406 }, { "epoch": 0.2726054922973878, "grad_norm": 4.281783086636439, "learning_rate": 3.403010033444816e-06, "loss": 2.023, "step": 407 }, { "epoch": 0.2732752846617549, "grad_norm": 2.5134651818030083, "learning_rate": 3.4113712374581942e-06, "loss": 2.0533, "step": 408 }, { "epoch": 0.2739450770261219, "grad_norm": 4.247544836877111, "learning_rate": 3.419732441471572e-06, "loss": 1.768, "step": 409 }, { "epoch": 0.27461486939048896, "grad_norm": 2.97378089860524, "learning_rate": 3.42809364548495e-06, "loss": 1.9353, "step": 410 }, { "epoch": 0.275284661754856, "grad_norm": 2.526136928768654, "learning_rate": 3.436454849498328e-06, "loss": 1.7178, "step": 411 }, { "epoch": 0.27595445411922304, "grad_norm": 2.52522601811223, "learning_rate": 3.444816053511706e-06, "loss": 1.8154, "step": 412 }, { "epoch": 0.2766242464835901, "grad_norm": 3.39341795237595, "learning_rate": 3.453177257525084e-06, "loss": 1.8713, "step": 413 }, { "epoch": 0.2772940388479571, "grad_norm": 3.5645786635119627, "learning_rate": 3.4615384615384617e-06, "loss": 1.886, "step": 414 }, { "epoch": 0.27796383121232415, "grad_norm": 3.325084507489375, "learning_rate": 3.46989966555184e-06, "loss": 1.8163, "step": 415 }, { "epoch": 0.27863362357669125, "grad_norm": 2.5627029073830805, "learning_rate": 3.4782608695652175e-06, "loss": 2.0298, "step": 416 }, { "epoch": 0.2793034159410583, "grad_norm": 2.479700893382351, "learning_rate": 3.4866220735785956e-06, "loss": 1.9256, "step": 417 }, { "epoch": 0.2799732083054253, "grad_norm": 3.728524624927344, "learning_rate": 3.4949832775919733e-06, "loss": 1.8193, "step": 418 }, { "epoch": 0.28064300066979236, "grad_norm": 3.1374836101634767, "learning_rate": 3.5033444816053515e-06, "loss": 1.8691, "step": 419 }, { "epoch": 0.2813127930341594, "grad_norm": 2.6675678790531827, "learning_rate": 3.511705685618729e-06, "loss": 1.8421, "step": 420 }, { "epoch": 0.28198258539852644, "grad_norm": 3.2935302533691324, "learning_rate": 3.5200668896321073e-06, "loss": 1.6928, "step": 421 }, { "epoch": 0.2826523777628935, "grad_norm": 7.331909469928724, "learning_rate": 3.528428093645485e-06, "loss": 1.7405, "step": 422 }, { "epoch": 0.2833221701272606, "grad_norm": 2.4756086952875918, "learning_rate": 3.536789297658863e-06, "loss": 1.8843, "step": 423 }, { "epoch": 0.2839919624916276, "grad_norm": 3.5922471608993765, "learning_rate": 3.5451505016722408e-06, "loss": 1.9451, "step": 424 }, { "epoch": 0.28466175485599465, "grad_norm": 2.909363683318659, "learning_rate": 3.553511705685619e-06, "loss": 1.8408, "step": 425 }, { "epoch": 0.2853315472203617, "grad_norm": 3.0711409773764373, "learning_rate": 3.5618729096989966e-06, "loss": 1.8605, "step": 426 }, { "epoch": 0.28600133958472873, "grad_norm": 2.4092567512116996, "learning_rate": 3.5702341137123747e-06, "loss": 1.7658, "step": 427 }, { "epoch": 0.28667113194909577, "grad_norm": 3.1288441850434854, "learning_rate": 3.578595317725753e-06, "loss": 1.7632, "step": 428 }, { "epoch": 0.2873409243134628, "grad_norm": 2.387124127508676, "learning_rate": 3.5869565217391305e-06, "loss": 1.9557, "step": 429 }, { "epoch": 0.28801071667782985, "grad_norm": 6.975320200645165, "learning_rate": 3.5953177257525087e-06, "loss": 1.9328, "step": 430 }, { "epoch": 0.28868050904219694, "grad_norm": 3.7466248654576138, "learning_rate": 3.6036789297658864e-06, "loss": 1.964, "step": 431 }, { "epoch": 0.289350301406564, "grad_norm": 2.380454766268421, "learning_rate": 3.6120401337792645e-06, "loss": 2.0197, "step": 432 }, { "epoch": 0.290020093770931, "grad_norm": 2.2292514879860454, "learning_rate": 3.620401337792642e-06, "loss": 1.8044, "step": 433 }, { "epoch": 0.29068988613529806, "grad_norm": 2.362977881669947, "learning_rate": 3.6287625418060203e-06, "loss": 1.8267, "step": 434 }, { "epoch": 0.2913596784996651, "grad_norm": 2.920392854578652, "learning_rate": 3.637123745819398e-06, "loss": 1.9178, "step": 435 }, { "epoch": 0.29202947086403214, "grad_norm": 6.400704242696343, "learning_rate": 3.645484949832776e-06, "loss": 1.5167, "step": 436 }, { "epoch": 0.2926992632283992, "grad_norm": 3.3841209720314622, "learning_rate": 3.653846153846154e-06, "loss": 1.8417, "step": 437 }, { "epoch": 0.29336905559276627, "grad_norm": 2.3361148573726727, "learning_rate": 3.662207357859532e-06, "loss": 1.9476, "step": 438 }, { "epoch": 0.2940388479571333, "grad_norm": 3.8148896152232594, "learning_rate": 3.6705685618729096e-06, "loss": 1.6744, "step": 439 }, { "epoch": 0.29470864032150035, "grad_norm": 2.2939220909295, "learning_rate": 3.6789297658862878e-06, "loss": 1.7012, "step": 440 }, { "epoch": 0.2953784326858674, "grad_norm": 3.252681560735929, "learning_rate": 3.6872909698996655e-06, "loss": 1.7035, "step": 441 }, { "epoch": 0.2960482250502344, "grad_norm": 5.264085831507128, "learning_rate": 3.6956521739130436e-06, "loss": 1.7137, "step": 442 }, { "epoch": 0.29671801741460146, "grad_norm": 4.085717547970395, "learning_rate": 3.7040133779264213e-06, "loss": 1.8724, "step": 443 }, { "epoch": 0.2973878097789685, "grad_norm": 2.6396419444023125, "learning_rate": 3.7123745819397994e-06, "loss": 1.6144, "step": 444 }, { "epoch": 0.29805760214333554, "grad_norm": 2.48738323445996, "learning_rate": 3.720735785953177e-06, "loss": 1.6931, "step": 445 }, { "epoch": 0.29872739450770264, "grad_norm": 2.4856199670871533, "learning_rate": 3.7290969899665552e-06, "loss": 1.9556, "step": 446 }, { "epoch": 0.2993971868720697, "grad_norm": 13.273555913364566, "learning_rate": 3.7374581939799333e-06, "loss": 1.8124, "step": 447 }, { "epoch": 0.3000669792364367, "grad_norm": 6.204214474317941, "learning_rate": 3.745819397993311e-06, "loss": 1.9334, "step": 448 }, { "epoch": 0.30073677160080375, "grad_norm": 5.565873222994077, "learning_rate": 3.7541806020066896e-06, "loss": 1.9187, "step": 449 }, { "epoch": 0.3014065639651708, "grad_norm": 2.6331896689297856, "learning_rate": 3.7625418060200673e-06, "loss": 1.8545, "step": 450 }, { "epoch": 0.30207635632953783, "grad_norm": 2.2790136479536023, "learning_rate": 3.7709030100334454e-06, "loss": 1.7348, "step": 451 }, { "epoch": 0.30274614869390487, "grad_norm": 2.5815876148453696, "learning_rate": 3.7792642140468235e-06, "loss": 1.7397, "step": 452 }, { "epoch": 0.3034159410582719, "grad_norm": 5.4160918829989715, "learning_rate": 3.7876254180602012e-06, "loss": 1.9373, "step": 453 }, { "epoch": 0.304085733422639, "grad_norm": 2.5642059092575002, "learning_rate": 3.7959866220735793e-06, "loss": 1.869, "step": 454 }, { "epoch": 0.30475552578700604, "grad_norm": 6.078189084808482, "learning_rate": 3.804347826086957e-06, "loss": 1.7488, "step": 455 }, { "epoch": 0.3054253181513731, "grad_norm": 2.271091887708997, "learning_rate": 3.812709030100335e-06, "loss": 1.7798, "step": 456 }, { "epoch": 0.3060951105157401, "grad_norm": 22.183593679377353, "learning_rate": 3.821070234113713e-06, "loss": 1.6855, "step": 457 }, { "epoch": 0.30676490288010716, "grad_norm": 2.487095588115721, "learning_rate": 3.8294314381270906e-06, "loss": 1.7869, "step": 458 }, { "epoch": 0.3074346952444742, "grad_norm": 2.3074375588915146, "learning_rate": 3.837792642140469e-06, "loss": 2.01, "step": 459 }, { "epoch": 0.30810448760884124, "grad_norm": 4.462132232091161, "learning_rate": 3.846153846153847e-06, "loss": 1.898, "step": 460 }, { "epoch": 0.30877427997320833, "grad_norm": 3.550050114862481, "learning_rate": 3.854515050167225e-06, "loss": 1.7651, "step": 461 }, { "epoch": 0.30944407233757537, "grad_norm": 3.48120020668745, "learning_rate": 3.862876254180602e-06, "loss": 1.7925, "step": 462 }, { "epoch": 0.3101138647019424, "grad_norm": 2.216673357242311, "learning_rate": 3.87123745819398e-06, "loss": 1.9674, "step": 463 }, { "epoch": 0.31078365706630945, "grad_norm": 3.0630422234088694, "learning_rate": 3.8795986622073584e-06, "loss": 1.8076, "step": 464 }, { "epoch": 0.3114534494306765, "grad_norm": 2.537908180199518, "learning_rate": 3.8879598662207366e-06, "loss": 1.9081, "step": 465 }, { "epoch": 0.3121232417950435, "grad_norm": 2.3609685122908513, "learning_rate": 3.896321070234114e-06, "loss": 1.9666, "step": 466 }, { "epoch": 0.31279303415941057, "grad_norm": 2.2388875475822534, "learning_rate": 3.904682274247492e-06, "loss": 1.681, "step": 467 }, { "epoch": 0.3134628265237776, "grad_norm": 2.478506941148036, "learning_rate": 3.91304347826087e-06, "loss": 1.8847, "step": 468 }, { "epoch": 0.3141326188881447, "grad_norm": 2.2607072769213477, "learning_rate": 3.921404682274248e-06, "loss": 1.8344, "step": 469 }, { "epoch": 0.31480241125251174, "grad_norm": 2.4300608056248447, "learning_rate": 3.929765886287626e-06, "loss": 1.6042, "step": 470 }, { "epoch": 0.3154722036168788, "grad_norm": 4.031658398716496, "learning_rate": 3.938127090301004e-06, "loss": 1.7492, "step": 471 }, { "epoch": 0.3161419959812458, "grad_norm": 2.4389466912122146, "learning_rate": 3.946488294314382e-06, "loss": 1.7934, "step": 472 }, { "epoch": 0.31681178834561285, "grad_norm": 2.2859329868060256, "learning_rate": 3.95484949832776e-06, "loss": 1.8343, "step": 473 }, { "epoch": 0.3174815807099799, "grad_norm": 2.395156102416765, "learning_rate": 3.963210702341138e-06, "loss": 1.8753, "step": 474 }, { "epoch": 0.31815137307434693, "grad_norm": 2.470945700673351, "learning_rate": 3.971571906354515e-06, "loss": 1.7653, "step": 475 }, { "epoch": 0.31882116543871397, "grad_norm": 4.094810456371656, "learning_rate": 3.979933110367893e-06, "loss": 1.9043, "step": 476 }, { "epoch": 0.31949095780308107, "grad_norm": 6.467119195235829, "learning_rate": 3.9882943143812715e-06, "loss": 1.6877, "step": 477 }, { "epoch": 0.3201607501674481, "grad_norm": 3.484364697526492, "learning_rate": 3.99665551839465e-06, "loss": 1.5972, "step": 478 }, { "epoch": 0.32083054253181514, "grad_norm": 2.4956509491110137, "learning_rate": 4.005016722408027e-06, "loss": 1.9498, "step": 479 }, { "epoch": 0.3215003348961822, "grad_norm": 3.4114734996609615, "learning_rate": 4.013377926421405e-06, "loss": 1.807, "step": 480 }, { "epoch": 0.3221701272605492, "grad_norm": 5.98152999202746, "learning_rate": 4.021739130434783e-06, "loss": 1.947, "step": 481 }, { "epoch": 0.32283991962491626, "grad_norm": 4.95580275157947, "learning_rate": 4.030100334448161e-06, "loss": 1.7282, "step": 482 }, { "epoch": 0.3235097119892833, "grad_norm": 2.62112957777427, "learning_rate": 4.0384615384615385e-06, "loss": 1.8168, "step": 483 }, { "epoch": 0.3241795043536504, "grad_norm": 4.172874960026993, "learning_rate": 4.046822742474917e-06, "loss": 1.985, "step": 484 }, { "epoch": 0.32484929671801743, "grad_norm": 2.576483927050478, "learning_rate": 4.055183946488295e-06, "loss": 1.8439, "step": 485 }, { "epoch": 0.32551908908238447, "grad_norm": 2.565628618931384, "learning_rate": 4.063545150501673e-06, "loss": 1.616, "step": 486 }, { "epoch": 0.3261888814467515, "grad_norm": 4.815118937232349, "learning_rate": 4.071906354515051e-06, "loss": 2.033, "step": 487 }, { "epoch": 0.32685867381111855, "grad_norm": 11.777843344417775, "learning_rate": 4.080267558528428e-06, "loss": 1.7509, "step": 488 }, { "epoch": 0.3275284661754856, "grad_norm": 4.044958150870569, "learning_rate": 4.088628762541806e-06, "loss": 2.1124, "step": 489 }, { "epoch": 0.3281982585398526, "grad_norm": 2.8055021520680863, "learning_rate": 4.0969899665551845e-06, "loss": 2.0182, "step": 490 }, { "epoch": 0.32886805090421967, "grad_norm": 2.6639444288029996, "learning_rate": 4.105351170568563e-06, "loss": 1.9199, "step": 491 }, { "epoch": 0.32953784326858676, "grad_norm": 2.741052914212561, "learning_rate": 4.11371237458194e-06, "loss": 1.7402, "step": 492 }, { "epoch": 0.3302076356329538, "grad_norm": 2.5182269952192224, "learning_rate": 4.122073578595318e-06, "loss": 1.8445, "step": 493 }, { "epoch": 0.33087742799732084, "grad_norm": 2.3856449737835286, "learning_rate": 4.130434782608696e-06, "loss": 1.8654, "step": 494 }, { "epoch": 0.3315472203616879, "grad_norm": 2.7067138475150143, "learning_rate": 4.138795986622074e-06, "loss": 1.6721, "step": 495 }, { "epoch": 0.3322170127260549, "grad_norm": 2.5298602962750008, "learning_rate": 4.1471571906354515e-06, "loss": 1.9663, "step": 496 }, { "epoch": 0.33288680509042196, "grad_norm": 2.6435994519345973, "learning_rate": 4.15551839464883e-06, "loss": 2.0223, "step": 497 }, { "epoch": 0.333556597454789, "grad_norm": 2.6016412665969697, "learning_rate": 4.163879598662208e-06, "loss": 1.7613, "step": 498 }, { "epoch": 0.3342263898191561, "grad_norm": 7.389169910997715, "learning_rate": 4.172240802675586e-06, "loss": 1.9241, "step": 499 }, { "epoch": 0.33489618218352313, "grad_norm": 2.8603314605879464, "learning_rate": 4.180602006688963e-06, "loss": 1.8539, "step": 500 }, { "epoch": 0.33556597454789017, "grad_norm": 4.128456652674315, "learning_rate": 4.188963210702341e-06, "loss": 1.7437, "step": 501 }, { "epoch": 0.3362357669122572, "grad_norm": 4.588045678225762, "learning_rate": 4.197324414715719e-06, "loss": 2.0697, "step": 502 }, { "epoch": 0.33690555927662424, "grad_norm": 2.7601911275658013, "learning_rate": 4.2056856187290975e-06, "loss": 2.0135, "step": 503 }, { "epoch": 0.3375753516409913, "grad_norm": 2.3015037476304476, "learning_rate": 4.214046822742475e-06, "loss": 1.9656, "step": 504 }, { "epoch": 0.3382451440053583, "grad_norm": 3.5982575817952145, "learning_rate": 4.222408026755853e-06, "loss": 1.9191, "step": 505 }, { "epoch": 0.33891493636972536, "grad_norm": 3.91531124938923, "learning_rate": 4.230769230769231e-06, "loss": 1.8929, "step": 506 }, { "epoch": 0.33958472873409246, "grad_norm": 4.788580399107024, "learning_rate": 4.239130434782609e-06, "loss": 1.7479, "step": 507 }, { "epoch": 0.3402545210984595, "grad_norm": 2.5867133015492736, "learning_rate": 4.247491638795987e-06, "loss": 1.8911, "step": 508 }, { "epoch": 0.34092431346282653, "grad_norm": 2.1307492241904917, "learning_rate": 4.2558528428093646e-06, "loss": 1.856, "step": 509 }, { "epoch": 0.3415941058271936, "grad_norm": 2.494155295115532, "learning_rate": 4.264214046822743e-06, "loss": 1.7429, "step": 510 }, { "epoch": 0.3422638981915606, "grad_norm": 2.2591962188702293, "learning_rate": 4.272575250836121e-06, "loss": 1.9171, "step": 511 }, { "epoch": 0.34293369055592765, "grad_norm": 3.947052192704875, "learning_rate": 4.280936454849499e-06, "loss": 1.747, "step": 512 }, { "epoch": 0.3436034829202947, "grad_norm": 2.685820979419144, "learning_rate": 4.289297658862876e-06, "loss": 2.0395, "step": 513 }, { "epoch": 0.34427327528466173, "grad_norm": 3.6982317551581163, "learning_rate": 4.297658862876254e-06, "loss": 1.9921, "step": 514 }, { "epoch": 0.3449430676490288, "grad_norm": 2.3260043427402963, "learning_rate": 4.3060200668896324e-06, "loss": 1.8022, "step": 515 }, { "epoch": 0.34561286001339586, "grad_norm": 2.84833891832341, "learning_rate": 4.3143812709030106e-06, "loss": 1.7879, "step": 516 }, { "epoch": 0.3462826523777629, "grad_norm": 2.375755388249524, "learning_rate": 4.322742474916388e-06, "loss": 1.863, "step": 517 }, { "epoch": 0.34695244474212994, "grad_norm": 3.544275849468136, "learning_rate": 4.331103678929766e-06, "loss": 1.7867, "step": 518 }, { "epoch": 0.347622237106497, "grad_norm": 2.37389823854014, "learning_rate": 4.339464882943144e-06, "loss": 1.7502, "step": 519 }, { "epoch": 0.348292029470864, "grad_norm": 2.5574812446873447, "learning_rate": 4.347826086956522e-06, "loss": 1.5025, "step": 520 }, { "epoch": 0.34896182183523106, "grad_norm": 3.661192306443199, "learning_rate": 4.3561872909698995e-06, "loss": 1.5008, "step": 521 }, { "epoch": 0.34963161419959815, "grad_norm": 4.285919484900424, "learning_rate": 4.364548494983278e-06, "loss": 1.8834, "step": 522 }, { "epoch": 0.3503014065639652, "grad_norm": 2.438008439377152, "learning_rate": 4.372909698996656e-06, "loss": 1.9387, "step": 523 }, { "epoch": 0.35097119892833223, "grad_norm": 2.1782887864580003, "learning_rate": 4.381270903010034e-06, "loss": 1.7859, "step": 524 }, { "epoch": 0.35164099129269927, "grad_norm": 12.632253767365633, "learning_rate": 4.389632107023412e-06, "loss": 1.6249, "step": 525 }, { "epoch": 0.3523107836570663, "grad_norm": 3.448386098039535, "learning_rate": 4.397993311036789e-06, "loss": 1.805, "step": 526 }, { "epoch": 0.35298057602143335, "grad_norm": 3.7076397863645605, "learning_rate": 4.406354515050167e-06, "loss": 1.8104, "step": 527 }, { "epoch": 0.3536503683858004, "grad_norm": 3.969063401418986, "learning_rate": 4.4147157190635455e-06, "loss": 2.0012, "step": 528 }, { "epoch": 0.3543201607501674, "grad_norm": 3.3131296152821275, "learning_rate": 4.423076923076924e-06, "loss": 1.8193, "step": 529 }, { "epoch": 0.3549899531145345, "grad_norm": 2.1666702227711205, "learning_rate": 4.431438127090301e-06, "loss": 1.9356, "step": 530 }, { "epoch": 0.35565974547890156, "grad_norm": 2.214264840384012, "learning_rate": 4.439799331103679e-06, "loss": 1.683, "step": 531 }, { "epoch": 0.3563295378432686, "grad_norm": 2.2327694121091164, "learning_rate": 4.448160535117057e-06, "loss": 1.8345, "step": 532 }, { "epoch": 0.35699933020763563, "grad_norm": 2.211384270328829, "learning_rate": 4.456521739130435e-06, "loss": 1.835, "step": 533 }, { "epoch": 0.3576691225720027, "grad_norm": 2.748388651374809, "learning_rate": 4.4648829431438125e-06, "loss": 1.8267, "step": 534 }, { "epoch": 0.3583389149363697, "grad_norm": 3.0459541821417457, "learning_rate": 4.473244147157191e-06, "loss": 2.0196, "step": 535 }, { "epoch": 0.35900870730073675, "grad_norm": 3.6517476951069785, "learning_rate": 4.481605351170569e-06, "loss": 1.8042, "step": 536 }, { "epoch": 0.3596784996651038, "grad_norm": 2.5055995556435073, "learning_rate": 4.489966555183947e-06, "loss": 1.9178, "step": 537 }, { "epoch": 0.3603482920294709, "grad_norm": 2.5184996206915997, "learning_rate": 4.498327759197324e-06, "loss": 1.8516, "step": 538 }, { "epoch": 0.3610180843938379, "grad_norm": 4.9212285669560165, "learning_rate": 4.506688963210702e-06, "loss": 1.6814, "step": 539 }, { "epoch": 0.36168787675820496, "grad_norm": 3.6047791593626934, "learning_rate": 4.51505016722408e-06, "loss": 1.6965, "step": 540 }, { "epoch": 0.362357669122572, "grad_norm": 5.026130898222163, "learning_rate": 4.5234113712374585e-06, "loss": 1.9075, "step": 541 }, { "epoch": 0.36302746148693904, "grad_norm": 2.3023410956467067, "learning_rate": 4.531772575250837e-06, "loss": 1.9425, "step": 542 }, { "epoch": 0.3636972538513061, "grad_norm": 2.516251547532507, "learning_rate": 4.540133779264214e-06, "loss": 1.9695, "step": 543 }, { "epoch": 0.3643670462156731, "grad_norm": 5.676513692588689, "learning_rate": 4.548494983277592e-06, "loss": 1.7292, "step": 544 }, { "epoch": 0.3650368385800402, "grad_norm": 4.944271781942838, "learning_rate": 4.55685618729097e-06, "loss": 1.8362, "step": 545 }, { "epoch": 0.36570663094440725, "grad_norm": 2.7517687489257656, "learning_rate": 4.565217391304348e-06, "loss": 1.5454, "step": 546 }, { "epoch": 0.3663764233087743, "grad_norm": 3.8764945756324685, "learning_rate": 4.5735785953177255e-06, "loss": 1.7493, "step": 547 }, { "epoch": 0.36704621567314133, "grad_norm": 2.2609782231975637, "learning_rate": 4.581939799331104e-06, "loss": 1.7817, "step": 548 }, { "epoch": 0.36771600803750837, "grad_norm": 2.484691735869524, "learning_rate": 4.590301003344483e-06, "loss": 2.0122, "step": 549 }, { "epoch": 0.3683858004018754, "grad_norm": 3.015341515506701, "learning_rate": 4.59866220735786e-06, "loss": 1.8463, "step": 550 }, { "epoch": 0.36905559276624245, "grad_norm": 2.20622838644935, "learning_rate": 4.607023411371238e-06, "loss": 1.8667, "step": 551 }, { "epoch": 0.3697253851306095, "grad_norm": 2.8027080158746696, "learning_rate": 4.615384615384616e-06, "loss": 1.9616, "step": 552 }, { "epoch": 0.3703951774949766, "grad_norm": 3.181753810669579, "learning_rate": 4.623745819397994e-06, "loss": 1.9655, "step": 553 }, { "epoch": 0.3710649698593436, "grad_norm": 4.497072092232785, "learning_rate": 4.6321070234113715e-06, "loss": 1.9239, "step": 554 }, { "epoch": 0.37173476222371066, "grad_norm": 3.8146378710053295, "learning_rate": 4.64046822742475e-06, "loss": 1.8494, "step": 555 }, { "epoch": 0.3724045545880777, "grad_norm": 2.2191969425896705, "learning_rate": 4.648829431438128e-06, "loss": 1.806, "step": 556 }, { "epoch": 0.37307434695244474, "grad_norm": 2.8542521057112005, "learning_rate": 4.657190635451506e-06, "loss": 1.5843, "step": 557 }, { "epoch": 0.3737441393168118, "grad_norm": 3.3281972169304788, "learning_rate": 4.665551839464883e-06, "loss": 1.9255, "step": 558 }, { "epoch": 0.3744139316811788, "grad_norm": 2.496154445784602, "learning_rate": 4.673913043478261e-06, "loss": 1.7646, "step": 559 }, { "epoch": 0.3750837240455459, "grad_norm": 2.531996548365815, "learning_rate": 4.6822742474916394e-06, "loss": 1.6594, "step": 560 }, { "epoch": 0.37575351640991295, "grad_norm": 3.1201092242526745, "learning_rate": 4.6906354515050175e-06, "loss": 1.8618, "step": 561 }, { "epoch": 0.37642330877428, "grad_norm": 2.359168364706417, "learning_rate": 4.698996655518395e-06, "loss": 1.7629, "step": 562 }, { "epoch": 0.377093101138647, "grad_norm": 4.058619865581199, "learning_rate": 4.707357859531773e-06, "loss": 2.1118, "step": 563 }, { "epoch": 0.37776289350301406, "grad_norm": 2.1710713478454133, "learning_rate": 4.715719063545151e-06, "loss": 1.7928, "step": 564 }, { "epoch": 0.3784326858673811, "grad_norm": 2.438233893136751, "learning_rate": 4.724080267558529e-06, "loss": 1.879, "step": 565 }, { "epoch": 0.37910247823174814, "grad_norm": 4.754536080549682, "learning_rate": 4.732441471571907e-06, "loss": 2.0681, "step": 566 }, { "epoch": 0.3797722705961152, "grad_norm": 2.187494343258754, "learning_rate": 4.740802675585285e-06, "loss": 1.9062, "step": 567 }, { "epoch": 0.3804420629604823, "grad_norm": 2.3635615857080188, "learning_rate": 4.749163879598663e-06, "loss": 1.9334, "step": 568 }, { "epoch": 0.3811118553248493, "grad_norm": 3.0577267604046496, "learning_rate": 4.757525083612041e-06, "loss": 1.8832, "step": 569 }, { "epoch": 0.38178164768921635, "grad_norm": 2.249068159954611, "learning_rate": 4.765886287625419e-06, "loss": 1.5995, "step": 570 }, { "epoch": 0.3824514400535834, "grad_norm": 4.603524775719262, "learning_rate": 4.774247491638796e-06, "loss": 1.7175, "step": 571 }, { "epoch": 0.38312123241795043, "grad_norm": 2.3013540892406072, "learning_rate": 4.782608695652174e-06, "loss": 1.9676, "step": 572 }, { "epoch": 0.38379102478231747, "grad_norm": 3.7598889835395513, "learning_rate": 4.7909698996655525e-06, "loss": 1.7107, "step": 573 }, { "epoch": 0.3844608171466845, "grad_norm": 2.2595230037413208, "learning_rate": 4.799331103678931e-06, "loss": 1.8146, "step": 574 }, { "epoch": 0.38513060951105155, "grad_norm": 4.5649996316334125, "learning_rate": 4.807692307692308e-06, "loss": 1.8798, "step": 575 }, { "epoch": 0.38580040187541864, "grad_norm": 3.38522531690688, "learning_rate": 4.816053511705686e-06, "loss": 1.8271, "step": 576 }, { "epoch": 0.3864701942397857, "grad_norm": 2.1512871856139864, "learning_rate": 4.824414715719064e-06, "loss": 1.7261, "step": 577 }, { "epoch": 0.3871399866041527, "grad_norm": 3.0687561219823785, "learning_rate": 4.832775919732442e-06, "loss": 1.7655, "step": 578 }, { "epoch": 0.38780977896851976, "grad_norm": 2.277444759676774, "learning_rate": 4.8411371237458195e-06, "loss": 1.6224, "step": 579 }, { "epoch": 0.3884795713328868, "grad_norm": 7.51716484464802, "learning_rate": 4.849498327759198e-06, "loss": 1.7413, "step": 580 }, { "epoch": 0.38914936369725384, "grad_norm": 3.309124100858334, "learning_rate": 4.857859531772576e-06, "loss": 1.6745, "step": 581 }, { "epoch": 0.3898191560616209, "grad_norm": 2.7352032431713584, "learning_rate": 4.866220735785954e-06, "loss": 1.6558, "step": 582 }, { "epoch": 0.39048894842598797, "grad_norm": 2.365142372569352, "learning_rate": 4.874581939799332e-06, "loss": 1.7374, "step": 583 }, { "epoch": 0.391158740790355, "grad_norm": 2.235072265820366, "learning_rate": 4.882943143812709e-06, "loss": 1.9054, "step": 584 }, { "epoch": 0.39182853315472205, "grad_norm": 2.8761530149416816, "learning_rate": 4.891304347826087e-06, "loss": 1.8917, "step": 585 }, { "epoch": 0.3924983255190891, "grad_norm": 2.221823722961997, "learning_rate": 4.8996655518394655e-06, "loss": 1.8974, "step": 586 }, { "epoch": 0.3931681178834561, "grad_norm": 2.4083264416472243, "learning_rate": 4.908026755852844e-06, "loss": 1.7683, "step": 587 }, { "epoch": 0.39383791024782316, "grad_norm": 2.227355499444258, "learning_rate": 4.916387959866221e-06, "loss": 1.8132, "step": 588 }, { "epoch": 0.3945077026121902, "grad_norm": 2.2734633052619837, "learning_rate": 4.924749163879599e-06, "loss": 1.9575, "step": 589 }, { "epoch": 0.39517749497655724, "grad_norm": 2.291968693774128, "learning_rate": 4.933110367892977e-06, "loss": 1.921, "step": 590 }, { "epoch": 0.39584728734092434, "grad_norm": 2.551500095674104, "learning_rate": 4.941471571906355e-06, "loss": 1.8591, "step": 591 }, { "epoch": 0.3965170797052914, "grad_norm": 3.438244942620692, "learning_rate": 4.9498327759197325e-06, "loss": 1.8648, "step": 592 }, { "epoch": 0.3971868720696584, "grad_norm": 4.6927297792718665, "learning_rate": 4.958193979933111e-06, "loss": 1.679, "step": 593 }, { "epoch": 0.39785666443402545, "grad_norm": 2.4711475345153056, "learning_rate": 4.966555183946489e-06, "loss": 1.7086, "step": 594 }, { "epoch": 0.3985264567983925, "grad_norm": 4.55790373241129, "learning_rate": 4.974916387959867e-06, "loss": 1.8259, "step": 595 }, { "epoch": 0.39919624916275953, "grad_norm": 2.3713811335909782, "learning_rate": 4.983277591973244e-06, "loss": 1.7712, "step": 596 }, { "epoch": 0.39986604152712657, "grad_norm": 2.1842822918294416, "learning_rate": 4.991638795986622e-06, "loss": 1.7928, "step": 597 }, { "epoch": 0.40053583389149366, "grad_norm": 2.325151136188292, "learning_rate": 5e-06, "loss": 2.0297, "step": 598 }, { "epoch": 0.4012056262558607, "grad_norm": 2.2842084480409297, "learning_rate": 4.9999995728165475e-06, "loss": 1.8496, "step": 599 }, { "epoch": 0.40187541862022774, "grad_norm": 3.6669266679673, "learning_rate": 4.9999982912663345e-06, "loss": 1.8415, "step": 600 }, { "epoch": 0.4025452109845948, "grad_norm": 4.719624494838494, "learning_rate": 4.9999961553498e-06, "loss": 1.7732, "step": 601 }, { "epoch": 0.4032150033489618, "grad_norm": 2.8362491033697355, "learning_rate": 4.999993165067674e-06, "loss": 1.6995, "step": 602 }, { "epoch": 0.40388479571332886, "grad_norm": 5.620204715693014, "learning_rate": 4.999989320420977e-06, "loss": 1.7679, "step": 603 }, { "epoch": 0.4045545880776959, "grad_norm": 4.5298432770084505, "learning_rate": 4.999984621411024e-06, "loss": 1.8657, "step": 604 }, { "epoch": 0.40522438044206294, "grad_norm": 5.240354638986383, "learning_rate": 4.99997906803942e-06, "loss": 1.867, "step": 605 }, { "epoch": 0.40589417280643003, "grad_norm": 3.348397594445817, "learning_rate": 4.999972660308064e-06, "loss": 1.8349, "step": 606 }, { "epoch": 0.40656396517079707, "grad_norm": 2.3024036586931453, "learning_rate": 4.9999653982191465e-06, "loss": 1.8963, "step": 607 }, { "epoch": 0.4072337575351641, "grad_norm": 3.5775841843467044, "learning_rate": 4.999957281775148e-06, "loss": 2.0266, "step": 608 }, { "epoch": 0.40790354989953115, "grad_norm": 2.2868633230645443, "learning_rate": 4.999948310978842e-06, "loss": 1.7335, "step": 609 }, { "epoch": 0.4085733422638982, "grad_norm": 4.194121156083787, "learning_rate": 4.999938485833293e-06, "loss": 1.8242, "step": 610 }, { "epoch": 0.4092431346282652, "grad_norm": 2.3737760403938717, "learning_rate": 4.999927806341863e-06, "loss": 1.8086, "step": 611 }, { "epoch": 0.40991292699263226, "grad_norm": 2.092246474264347, "learning_rate": 4.999916272508198e-06, "loss": 1.845, "step": 612 }, { "epoch": 0.4105827193569993, "grad_norm": 5.636196846684404, "learning_rate": 4.99990388433624e-06, "loss": 1.7833, "step": 613 }, { "epoch": 0.4112525117213664, "grad_norm": 3.5096129004314234, "learning_rate": 4.999890641830225e-06, "loss": 1.6131, "step": 614 }, { "epoch": 0.41192230408573344, "grad_norm": 2.494879611399409, "learning_rate": 4.999876544994676e-06, "loss": 1.7477, "step": 615 }, { "epoch": 0.4125920964501005, "grad_norm": 3.550068787559106, "learning_rate": 4.9998615938344106e-06, "loss": 1.7021, "step": 616 }, { "epoch": 0.4132618888144675, "grad_norm": 2.4833768673226615, "learning_rate": 4.9998457883545405e-06, "loss": 1.6179, "step": 617 }, { "epoch": 0.41393168117883455, "grad_norm": 2.363850994387893, "learning_rate": 4.999829128560465e-06, "loss": 1.8147, "step": 618 }, { "epoch": 0.4146014735432016, "grad_norm": 5.6799261319592755, "learning_rate": 4.999811614457879e-06, "loss": 1.756, "step": 619 }, { "epoch": 0.41527126590756863, "grad_norm": 2.2961031038823316, "learning_rate": 4.999793246052768e-06, "loss": 1.8893, "step": 620 }, { "epoch": 0.4159410582719357, "grad_norm": 2.2431505573623065, "learning_rate": 4.999774023351407e-06, "loss": 1.9564, "step": 621 }, { "epoch": 0.41661085063630277, "grad_norm": 2.97207173138461, "learning_rate": 4.999753946360368e-06, "loss": 1.7981, "step": 622 }, { "epoch": 0.4172806430006698, "grad_norm": 2.4118284987592884, "learning_rate": 4.999733015086511e-06, "loss": 1.7297, "step": 623 }, { "epoch": 0.41795043536503684, "grad_norm": 4.034255812935671, "learning_rate": 4.999711229536989e-06, "loss": 1.9882, "step": 624 }, { "epoch": 0.4186202277294039, "grad_norm": 2.3357912824118, "learning_rate": 4.999688589719248e-06, "loss": 1.6517, "step": 625 }, { "epoch": 0.4192900200937709, "grad_norm": 2.1357563775421458, "learning_rate": 4.999665095641024e-06, "loss": 1.7465, "step": 626 }, { "epoch": 0.41995981245813796, "grad_norm": 3.38382411566755, "learning_rate": 4.999640747310347e-06, "loss": 1.772, "step": 627 }, { "epoch": 0.420629604822505, "grad_norm": 2.216675118729019, "learning_rate": 4.999615544735537e-06, "loss": 1.9226, "step": 628 }, { "epoch": 0.4212993971868721, "grad_norm": 6.072965026383361, "learning_rate": 4.999589487925208e-06, "loss": 1.7511, "step": 629 }, { "epoch": 0.42196918955123913, "grad_norm": 4.9500769141322865, "learning_rate": 4.999562576888264e-06, "loss": 1.797, "step": 630 }, { "epoch": 0.42263898191560617, "grad_norm": 2.3245169324886588, "learning_rate": 4.999534811633903e-06, "loss": 1.8585, "step": 631 }, { "epoch": 0.4233087742799732, "grad_norm": 2.3704377056704264, "learning_rate": 4.999506192171611e-06, "loss": 1.8954, "step": 632 }, { "epoch": 0.42397856664434025, "grad_norm": 6.502795715493123, "learning_rate": 4.999476718511172e-06, "loss": 1.7004, "step": 633 }, { "epoch": 0.4246483590087073, "grad_norm": 2.8422931980404105, "learning_rate": 4.999446390662657e-06, "loss": 1.6218, "step": 634 }, { "epoch": 0.4253181513730743, "grad_norm": 3.742361280443987, "learning_rate": 4.999415208636429e-06, "loss": 1.9779, "step": 635 }, { "epoch": 0.42598794373744137, "grad_norm": 4.7087680464508095, "learning_rate": 4.999383172443146e-06, "loss": 1.8549, "step": 636 }, { "epoch": 0.42665773610180846, "grad_norm": 2.6557133631640366, "learning_rate": 4.999350282093757e-06, "loss": 1.8841, "step": 637 }, { "epoch": 0.4273275284661755, "grad_norm": 2.328067941353049, "learning_rate": 4.9993165375995e-06, "loss": 1.7449, "step": 638 }, { "epoch": 0.42799732083054254, "grad_norm": 2.2252019406231156, "learning_rate": 4.9992819389719085e-06, "loss": 1.8098, "step": 639 }, { "epoch": 0.4286671131949096, "grad_norm": 2.524285382014486, "learning_rate": 4.999246486222806e-06, "loss": 2.0632, "step": 640 }, { "epoch": 0.4293369055592766, "grad_norm": 2.1935811270420644, "learning_rate": 4.999210179364309e-06, "loss": 1.8499, "step": 641 }, { "epoch": 0.43000669792364365, "grad_norm": 2.1695349985912395, "learning_rate": 4.999173018408824e-06, "loss": 1.6772, "step": 642 }, { "epoch": 0.4306764902880107, "grad_norm": 2.598259865457582, "learning_rate": 4.999135003369052e-06, "loss": 1.707, "step": 643 }, { "epoch": 0.4313462826523778, "grad_norm": 4.706607072747362, "learning_rate": 4.999096134257984e-06, "loss": 1.7721, "step": 644 }, { "epoch": 0.4320160750167448, "grad_norm": 4.781325798010932, "learning_rate": 4.999056411088903e-06, "loss": 1.9241, "step": 645 }, { "epoch": 0.43268586738111187, "grad_norm": 5.584058312239832, "learning_rate": 4.999015833875385e-06, "loss": 1.8559, "step": 646 }, { "epoch": 0.4333556597454789, "grad_norm": 3.840459839143552, "learning_rate": 4.998974402631296e-06, "loss": 1.7282, "step": 647 }, { "epoch": 0.43402545210984594, "grad_norm": 4.3469340571744395, "learning_rate": 4.998932117370796e-06, "loss": 2.0441, "step": 648 }, { "epoch": 0.434695244474213, "grad_norm": 4.47254907671226, "learning_rate": 4.998888978108336e-06, "loss": 1.7673, "step": 649 }, { "epoch": 0.43536503683858, "grad_norm": 4.387823045012746, "learning_rate": 4.9988449848586585e-06, "loss": 1.8946, "step": 650 }, { "epoch": 0.43603482920294706, "grad_norm": 2.154892336035561, "learning_rate": 4.998800137636797e-06, "loss": 1.7545, "step": 651 }, { "epoch": 0.43670462156731416, "grad_norm": 3.2503659371378566, "learning_rate": 4.9987544364580794e-06, "loss": 1.672, "step": 652 }, { "epoch": 0.4373744139316812, "grad_norm": 2.628054516351069, "learning_rate": 4.998707881338123e-06, "loss": 1.8423, "step": 653 }, { "epoch": 0.43804420629604823, "grad_norm": 5.025288170358171, "learning_rate": 4.998660472292838e-06, "loss": 1.71, "step": 654 }, { "epoch": 0.43871399866041527, "grad_norm": 2.436353214142625, "learning_rate": 4.998612209338426e-06, "loss": 1.8511, "step": 655 }, { "epoch": 0.4393837910247823, "grad_norm": 2.3867113846352606, "learning_rate": 4.998563092491382e-06, "loss": 2.056, "step": 656 }, { "epoch": 0.44005358338914935, "grad_norm": 2.1655033622520694, "learning_rate": 4.998513121768489e-06, "loss": 1.7707, "step": 657 }, { "epoch": 0.4407233757535164, "grad_norm": 21.038427115826718, "learning_rate": 4.998462297186828e-06, "loss": 1.9056, "step": 658 }, { "epoch": 0.4413931681178835, "grad_norm": 3.1810790378494955, "learning_rate": 4.998410618763765e-06, "loss": 1.6695, "step": 659 }, { "epoch": 0.4420629604822505, "grad_norm": 4.575172282140339, "learning_rate": 4.998358086516962e-06, "loss": 1.9102, "step": 660 }, { "epoch": 0.44273275284661756, "grad_norm": 2.447931592257187, "learning_rate": 4.998304700464371e-06, "loss": 1.7903, "step": 661 }, { "epoch": 0.4434025452109846, "grad_norm": 3.3801097707817593, "learning_rate": 4.998250460624239e-06, "loss": 1.7102, "step": 662 }, { "epoch": 0.44407233757535164, "grad_norm": 4.310496405729879, "learning_rate": 4.9981953670151e-06, "loss": 1.8196, "step": 663 }, { "epoch": 0.4447421299397187, "grad_norm": 2.297121212915909, "learning_rate": 4.998139419655782e-06, "loss": 1.6356, "step": 664 }, { "epoch": 0.4454119223040857, "grad_norm": 2.3938403350795414, "learning_rate": 4.998082618565405e-06, "loss": 1.8963, "step": 665 }, { "epoch": 0.44608171466845276, "grad_norm": 2.3533899330327346, "learning_rate": 4.998024963763383e-06, "loss": 1.8384, "step": 666 }, { "epoch": 0.44675150703281985, "grad_norm": 2.8233557947937986, "learning_rate": 4.997966455269415e-06, "loss": 1.851, "step": 667 }, { "epoch": 0.4474212993971869, "grad_norm": 2.4251977278391204, "learning_rate": 4.9979070931035e-06, "loss": 1.7378, "step": 668 }, { "epoch": 0.44809109176155393, "grad_norm": 2.190002423253682, "learning_rate": 4.997846877285923e-06, "loss": 1.7914, "step": 669 }, { "epoch": 0.44876088412592097, "grad_norm": 4.490635596155279, "learning_rate": 4.997785807837263e-06, "loss": 1.7014, "step": 670 }, { "epoch": 0.449430676490288, "grad_norm": 3.7754307299442544, "learning_rate": 4.99772388477839e-06, "loss": 1.9855, "step": 671 }, { "epoch": 0.45010046885465504, "grad_norm": 2.019607207177449, "learning_rate": 4.9976611081304655e-06, "loss": 1.7211, "step": 672 }, { "epoch": 0.4507702612190221, "grad_norm": 2.1675415089775854, "learning_rate": 4.997597477914944e-06, "loss": 1.8949, "step": 673 }, { "epoch": 0.4514400535833891, "grad_norm": 2.939515616243093, "learning_rate": 4.997532994153571e-06, "loss": 1.7602, "step": 674 }, { "epoch": 0.4521098459477562, "grad_norm": 2.2328012041238128, "learning_rate": 4.997467656868384e-06, "loss": 1.8466, "step": 675 }, { "epoch": 0.45277963831212326, "grad_norm": 2.8089982464915066, "learning_rate": 4.9974014660817105e-06, "loss": 1.8003, "step": 676 }, { "epoch": 0.4534494306764903, "grad_norm": 3.2374578588938605, "learning_rate": 4.997334421816172e-06, "loss": 1.6594, "step": 677 }, { "epoch": 0.45411922304085733, "grad_norm": 2.500718470580677, "learning_rate": 4.99726652409468e-06, "loss": 1.6291, "step": 678 }, { "epoch": 0.4547890154052244, "grad_norm": 2.2208970680579014, "learning_rate": 4.997197772940439e-06, "loss": 1.7071, "step": 679 }, { "epoch": 0.4554588077695914, "grad_norm": 2.4185369595617336, "learning_rate": 4.9971281683769446e-06, "loss": 1.7054, "step": 680 }, { "epoch": 0.45612860013395845, "grad_norm": 4.966774048349189, "learning_rate": 4.997057710427983e-06, "loss": 1.9909, "step": 681 }, { "epoch": 0.45679839249832555, "grad_norm": 2.2091508698368427, "learning_rate": 4.996986399117633e-06, "loss": 1.759, "step": 682 }, { "epoch": 0.4574681848626926, "grad_norm": 3.7047824810093863, "learning_rate": 4.996914234470266e-06, "loss": 1.6691, "step": 683 }, { "epoch": 0.4581379772270596, "grad_norm": 2.2128586765213103, "learning_rate": 4.996841216510543e-06, "loss": 1.723, "step": 684 }, { "epoch": 0.45880776959142666, "grad_norm": 2.5112751905092683, "learning_rate": 4.996767345263419e-06, "loss": 1.9289, "step": 685 }, { "epoch": 0.4594775619557937, "grad_norm": 4.6648784033388, "learning_rate": 4.996692620754138e-06, "loss": 1.83, "step": 686 }, { "epoch": 0.46014735432016074, "grad_norm": 3.284052438343692, "learning_rate": 4.996617043008237e-06, "loss": 2.01, "step": 687 }, { "epoch": 0.4608171466845278, "grad_norm": 3.1119093447446975, "learning_rate": 4.996540612051544e-06, "loss": 1.8711, "step": 688 }, { "epoch": 0.4614869390488948, "grad_norm": 3.6460597210312287, "learning_rate": 4.996463327910182e-06, "loss": 1.8787, "step": 689 }, { "epoch": 0.4621567314132619, "grad_norm": 2.311936338631102, "learning_rate": 4.996385190610559e-06, "loss": 1.7603, "step": 690 }, { "epoch": 0.46282652377762895, "grad_norm": 2.231752205765466, "learning_rate": 4.996306200179379e-06, "loss": 1.8727, "step": 691 }, { "epoch": 0.463496316141996, "grad_norm": 2.474139810210444, "learning_rate": 4.996226356643638e-06, "loss": 1.9069, "step": 692 }, { "epoch": 0.46416610850636303, "grad_norm": 3.023189190635685, "learning_rate": 4.9961456600306215e-06, "loss": 1.5631, "step": 693 }, { "epoch": 0.46483590087073007, "grad_norm": 2.4931105517409904, "learning_rate": 4.996064110367908e-06, "loss": 1.8124, "step": 694 }, { "epoch": 0.4655056932350971, "grad_norm": 4.40718847121033, "learning_rate": 4.995981707683365e-06, "loss": 1.7833, "step": 695 }, { "epoch": 0.46617548559946415, "grad_norm": 27.377563115822227, "learning_rate": 4.995898452005155e-06, "loss": 1.7561, "step": 696 }, { "epoch": 0.46684527796383124, "grad_norm": 2.3012220161331323, "learning_rate": 4.99581434336173e-06, "loss": 1.7387, "step": 697 }, { "epoch": 0.4675150703281983, "grad_norm": 3.35245232534728, "learning_rate": 4.9957293817818335e-06, "loss": 1.8557, "step": 698 }, { "epoch": 0.4681848626925653, "grad_norm": 2.3710878305068177, "learning_rate": 4.995643567294502e-06, "loss": 1.9093, "step": 699 }, { "epoch": 0.46885465505693236, "grad_norm": 2.5492170288960465, "learning_rate": 4.995556899929059e-06, "loss": 1.8405, "step": 700 }, { "epoch": 0.4695244474212994, "grad_norm": 2.1692020408739396, "learning_rate": 4.995469379715128e-06, "loss": 1.9776, "step": 701 }, { "epoch": 0.47019423978566643, "grad_norm": 2.227598635969639, "learning_rate": 4.995381006682613e-06, "loss": 1.7914, "step": 702 }, { "epoch": 0.4708640321500335, "grad_norm": 2.6732640360668567, "learning_rate": 4.995291780861721e-06, "loss": 1.9347, "step": 703 }, { "epoch": 0.4715338245144005, "grad_norm": 2.3214095547390405, "learning_rate": 4.99520170228294e-06, "loss": 1.8213, "step": 704 }, { "epoch": 0.4722036168787676, "grad_norm": 2.1491426126868274, "learning_rate": 4.9951107709770565e-06, "loss": 1.7709, "step": 705 }, { "epoch": 0.47287340924313465, "grad_norm": 2.7261224563232838, "learning_rate": 4.9950189869751455e-06, "loss": 1.8431, "step": 706 }, { "epoch": 0.4735432016075017, "grad_norm": 4.573826737655657, "learning_rate": 4.9949263503085736e-06, "loss": 1.8026, "step": 707 }, { "epoch": 0.4742129939718687, "grad_norm": 2.194119542931565, "learning_rate": 4.994832861008999e-06, "loss": 1.8453, "step": 708 }, { "epoch": 0.47488278633623576, "grad_norm": 5.426517034952132, "learning_rate": 4.994738519108372e-06, "loss": 1.7613, "step": 709 }, { "epoch": 0.4755525787006028, "grad_norm": 3.816755620259893, "learning_rate": 4.994643324638933e-06, "loss": 1.9923, "step": 710 }, { "epoch": 0.47622237106496984, "grad_norm": 4.64125558471245, "learning_rate": 4.9945472776332135e-06, "loss": 1.8913, "step": 711 }, { "epoch": 0.4768921634293369, "grad_norm": 2.449067745682637, "learning_rate": 4.994450378124039e-06, "loss": 1.8144, "step": 712 }, { "epoch": 0.477561955793704, "grad_norm": 11.285731550106401, "learning_rate": 4.994352626144525e-06, "loss": 1.6342, "step": 713 }, { "epoch": 0.478231748158071, "grad_norm": 5.956206921299215, "learning_rate": 4.994254021728075e-06, "loss": 1.8531, "step": 714 }, { "epoch": 0.47890154052243805, "grad_norm": 2.066529763121893, "learning_rate": 4.99415456490839e-06, "loss": 1.8651, "step": 715 }, { "epoch": 0.4795713328868051, "grad_norm": 6.051874608908242, "learning_rate": 4.994054255719458e-06, "loss": 1.7669, "step": 716 }, { "epoch": 0.48024112525117213, "grad_norm": 2.5364308535706233, "learning_rate": 4.993953094195558e-06, "loss": 1.9337, "step": 717 }, { "epoch": 0.48091091761553917, "grad_norm": 3.94155252779343, "learning_rate": 4.993851080371262e-06, "loss": 1.8371, "step": 718 }, { "epoch": 0.4815807099799062, "grad_norm": 2.379804713065538, "learning_rate": 4.993748214281435e-06, "loss": 1.8018, "step": 719 }, { "epoch": 0.4822505023442733, "grad_norm": 2.0776397054078113, "learning_rate": 4.993644495961229e-06, "loss": 1.8848, "step": 720 }, { "epoch": 0.48292029470864034, "grad_norm": 10.118277507879771, "learning_rate": 4.99353992544609e-06, "loss": 1.7186, "step": 721 }, { "epoch": 0.4835900870730074, "grad_norm": 3.2466412874522907, "learning_rate": 4.993434502771755e-06, "loss": 1.8783, "step": 722 }, { "epoch": 0.4842598794373744, "grad_norm": 2.118268112277287, "learning_rate": 4.9933282279742515e-06, "loss": 1.9492, "step": 723 }, { "epoch": 0.48492967180174146, "grad_norm": 3.3757066576924846, "learning_rate": 4.993221101089898e-06, "loss": 1.8862, "step": 724 }, { "epoch": 0.4855994641661085, "grad_norm": 3.669100734895576, "learning_rate": 4.9931131221553065e-06, "loss": 1.6633, "step": 725 }, { "epoch": 0.48626925653047554, "grad_norm": 6.708193603923884, "learning_rate": 4.993004291207376e-06, "loss": 1.6992, "step": 726 }, { "epoch": 0.4869390488948426, "grad_norm": 6.263314602999536, "learning_rate": 4.992894608283301e-06, "loss": 1.8894, "step": 727 }, { "epoch": 0.48760884125920967, "grad_norm": 4.458286108402087, "learning_rate": 4.992784073420565e-06, "loss": 1.7643, "step": 728 }, { "epoch": 0.4882786336235767, "grad_norm": 2.3751715455651317, "learning_rate": 4.992672686656943e-06, "loss": 1.9251, "step": 729 }, { "epoch": 0.48894842598794375, "grad_norm": 4.568505511914387, "learning_rate": 4.9925604480305e-06, "loss": 1.6507, "step": 730 }, { "epoch": 0.4896182183523108, "grad_norm": 2.508089608333261, "learning_rate": 4.992447357579594e-06, "loss": 1.715, "step": 731 }, { "epoch": 0.4902880107166778, "grad_norm": 2.419418195657542, "learning_rate": 4.992333415342874e-06, "loss": 1.743, "step": 732 }, { "epoch": 0.49095780308104486, "grad_norm": 3.6049183905483986, "learning_rate": 4.992218621359278e-06, "loss": 1.6547, "step": 733 }, { "epoch": 0.4916275954454119, "grad_norm": 2.681459771436212, "learning_rate": 4.9921029756680374e-06, "loss": 1.7087, "step": 734 }, { "epoch": 0.49229738780977894, "grad_norm": 2.634901547600657, "learning_rate": 4.9919864783086725e-06, "loss": 1.832, "step": 735 }, { "epoch": 0.49296718017414604, "grad_norm": 2.226052289936058, "learning_rate": 4.991869129320997e-06, "loss": 1.7401, "step": 736 }, { "epoch": 0.4936369725385131, "grad_norm": 3.777614441830342, "learning_rate": 4.991750928745116e-06, "loss": 1.8887, "step": 737 }, { "epoch": 0.4943067649028801, "grad_norm": 2.471492671260381, "learning_rate": 4.991631876621421e-06, "loss": 1.8713, "step": 738 }, { "epoch": 0.49497655726724715, "grad_norm": 3.0899454060783773, "learning_rate": 4.9915119729906e-06, "loss": 1.7622, "step": 739 }, { "epoch": 0.4956463496316142, "grad_norm": 6.181375147245705, "learning_rate": 4.991391217893628e-06, "loss": 1.7696, "step": 740 }, { "epoch": 0.49631614199598123, "grad_norm": 3.313430600356463, "learning_rate": 4.991269611371774e-06, "loss": 1.8885, "step": 741 }, { "epoch": 0.49698593436034827, "grad_norm": 3.373822426938885, "learning_rate": 4.991147153466596e-06, "loss": 1.8447, "step": 742 }, { "epoch": 0.49765572672471536, "grad_norm": 2.335738333114462, "learning_rate": 4.991023844219945e-06, "loss": 1.8808, "step": 743 }, { "epoch": 0.4983255190890824, "grad_norm": 2.361833550822361, "learning_rate": 4.9908996836739584e-06, "loss": 1.7596, "step": 744 }, { "epoch": 0.49899531145344944, "grad_norm": 3.191619610881239, "learning_rate": 4.990774671871071e-06, "loss": 1.9309, "step": 745 }, { "epoch": 0.4996651038178165, "grad_norm": 2.1711170431844633, "learning_rate": 4.9906488088540025e-06, "loss": 1.6834, "step": 746 }, { "epoch": 0.5003348961821835, "grad_norm": 2.261642319114356, "learning_rate": 4.9905220946657685e-06, "loss": 1.807, "step": 747 }, { "epoch": 0.5010046885465506, "grad_norm": 2.176348835726146, "learning_rate": 4.990394529349672e-06, "loss": 1.7173, "step": 748 }, { "epoch": 0.5016744809109176, "grad_norm": 2.3454178628591404, "learning_rate": 4.990266112949307e-06, "loss": 1.9941, "step": 749 }, { "epoch": 0.5023442732752846, "grad_norm": 6.668531944020338, "learning_rate": 4.9901368455085616e-06, "loss": 1.5744, "step": 750 }, { "epoch": 0.5030140656396517, "grad_norm": 3.0343430797269235, "learning_rate": 4.990006727071612e-06, "loss": 1.8064, "step": 751 }, { "epoch": 0.5036838580040187, "grad_norm": 4.179828862362998, "learning_rate": 4.989875757682924e-06, "loss": 1.9414, "step": 752 }, { "epoch": 0.5043536503683858, "grad_norm": 2.484976399355909, "learning_rate": 4.989743937387257e-06, "loss": 1.7485, "step": 753 }, { "epoch": 0.5050234427327528, "grad_norm": 2.6818494106913677, "learning_rate": 4.989611266229662e-06, "loss": 1.5365, "step": 754 }, { "epoch": 0.5056932350971199, "grad_norm": 2.1514087101766664, "learning_rate": 4.989477744255475e-06, "loss": 1.9372, "step": 755 }, { "epoch": 0.506363027461487, "grad_norm": 2.4451124284703636, "learning_rate": 4.98934337151033e-06, "loss": 2.0597, "step": 756 }, { "epoch": 0.507032819825854, "grad_norm": 2.4155420395852385, "learning_rate": 4.989208148040148e-06, "loss": 1.8503, "step": 757 }, { "epoch": 0.5077026121902211, "grad_norm": 2.4577152397651614, "learning_rate": 4.9890720738911405e-06, "loss": 1.7006, "step": 758 }, { "epoch": 0.5083724045545881, "grad_norm": 4.334706112787799, "learning_rate": 4.9889351491098095e-06, "loss": 1.8037, "step": 759 }, { "epoch": 0.5090421969189551, "grad_norm": 2.240017273103996, "learning_rate": 4.9887973737429505e-06, "loss": 1.6984, "step": 760 }, { "epoch": 0.5097119892833222, "grad_norm": 2.342231708190007, "learning_rate": 4.988658747837647e-06, "loss": 1.7591, "step": 761 }, { "epoch": 0.5103817816476892, "grad_norm": 2.3113601454111166, "learning_rate": 4.988519271441272e-06, "loss": 1.9411, "step": 762 }, { "epoch": 0.5110515740120563, "grad_norm": 3.7169941481808633, "learning_rate": 4.988378944601495e-06, "loss": 1.758, "step": 763 }, { "epoch": 0.5117213663764233, "grad_norm": 3.5625843836826405, "learning_rate": 4.98823776736627e-06, "loss": 1.7589, "step": 764 }, { "epoch": 0.5123911587407903, "grad_norm": 2.5158276990915702, "learning_rate": 4.988095739783843e-06, "loss": 1.7734, "step": 765 }, { "epoch": 0.5130609511051574, "grad_norm": 2.1486462787353298, "learning_rate": 4.987952861902753e-06, "loss": 1.9857, "step": 766 }, { "epoch": 0.5137307434695244, "grad_norm": 3.398567004198017, "learning_rate": 4.987809133771828e-06, "loss": 1.8266, "step": 767 }, { "epoch": 0.5144005358338914, "grad_norm": 6.556949290632318, "learning_rate": 4.987664555440186e-06, "loss": 1.7823, "step": 768 }, { "epoch": 0.5150703281982585, "grad_norm": 4.879364241016928, "learning_rate": 4.9875191269572356e-06, "loss": 1.8179, "step": 769 }, { "epoch": 0.5157401205626256, "grad_norm": 5.259576106359075, "learning_rate": 4.987372848372678e-06, "loss": 1.7621, "step": 770 }, { "epoch": 0.5164099129269927, "grad_norm": 2.285229415903265, "learning_rate": 4.987225719736503e-06, "loss": 1.8689, "step": 771 }, { "epoch": 0.5170797052913597, "grad_norm": 2.5845491715691655, "learning_rate": 4.987077741098991e-06, "loss": 1.8107, "step": 772 }, { "epoch": 0.5177494976557268, "grad_norm": 8.54280558957128, "learning_rate": 4.986928912510712e-06, "loss": 1.8176, "step": 773 }, { "epoch": 0.5184192900200938, "grad_norm": 2.5329285879373953, "learning_rate": 4.98677923402253e-06, "loss": 1.8156, "step": 774 }, { "epoch": 0.5190890823844608, "grad_norm": 2.4546854277915657, "learning_rate": 4.986628705685597e-06, "loss": 1.8139, "step": 775 }, { "epoch": 0.5197588747488279, "grad_norm": 2.2568928537760673, "learning_rate": 4.986477327551353e-06, "loss": 1.74, "step": 776 }, { "epoch": 0.5204286671131949, "grad_norm": 2.418341117238603, "learning_rate": 4.986325099671534e-06, "loss": 1.7335, "step": 777 }, { "epoch": 0.521098459477562, "grad_norm": 2.3887817108413136, "learning_rate": 4.986172022098161e-06, "loss": 1.9166, "step": 778 }, { "epoch": 0.521768251841929, "grad_norm": 2.2979962602803483, "learning_rate": 4.986018094883549e-06, "loss": 1.8274, "step": 779 }, { "epoch": 0.522438044206296, "grad_norm": 3.2089571497848546, "learning_rate": 4.985863318080302e-06, "loss": 1.7014, "step": 780 }, { "epoch": 0.5231078365706631, "grad_norm": 2.414986434473166, "learning_rate": 4.985707691741315e-06, "loss": 1.7319, "step": 781 }, { "epoch": 0.5237776289350301, "grad_norm": 2.7978194820120597, "learning_rate": 4.985551215919772e-06, "loss": 1.7401, "step": 782 }, { "epoch": 0.5244474212993971, "grad_norm": 2.6979575723879465, "learning_rate": 4.985393890669148e-06, "loss": 1.7269, "step": 783 }, { "epoch": 0.5251172136637642, "grad_norm": 2.385968658255069, "learning_rate": 4.9852357160432085e-06, "loss": 1.8013, "step": 784 }, { "epoch": 0.5257870060281313, "grad_norm": 2.9302292081577472, "learning_rate": 4.98507669209601e-06, "loss": 1.6214, "step": 785 }, { "epoch": 0.5264567983924984, "grad_norm": 2.904749855719379, "learning_rate": 4.984916818881898e-06, "loss": 1.8412, "step": 786 }, { "epoch": 0.5271265907568654, "grad_norm": 2.544064139955138, "learning_rate": 4.984756096455507e-06, "loss": 1.6643, "step": 787 }, { "epoch": 0.5277963831212324, "grad_norm": 4.234298098525975, "learning_rate": 4.984594524871765e-06, "loss": 1.736, "step": 788 }, { "epoch": 0.5284661754855995, "grad_norm": 4.398176382474912, "learning_rate": 4.98443210418589e-06, "loss": 1.7907, "step": 789 }, { "epoch": 0.5291359678499665, "grad_norm": 2.5783522872240203, "learning_rate": 4.984268834453386e-06, "loss": 1.8209, "step": 790 }, { "epoch": 0.5298057602143336, "grad_norm": 2.209672675777014, "learning_rate": 4.9841047157300515e-06, "loss": 1.8362, "step": 791 }, { "epoch": 0.5304755525787006, "grad_norm": 9.735234631438914, "learning_rate": 4.983939748071972e-06, "loss": 1.8175, "step": 792 }, { "epoch": 0.5311453449430676, "grad_norm": 4.25874556718552, "learning_rate": 4.983773931535527e-06, "loss": 1.9348, "step": 793 }, { "epoch": 0.5318151373074347, "grad_norm": 2.4457107850319666, "learning_rate": 4.983607266177381e-06, "loss": 1.9585, "step": 794 }, { "epoch": 0.5324849296718017, "grad_norm": 2.1591273504198347, "learning_rate": 4.9834397520544945e-06, "loss": 1.8935, "step": 795 }, { "epoch": 0.5331547220361688, "grad_norm": 3.1154461396089657, "learning_rate": 4.983271389224113e-06, "loss": 1.7739, "step": 796 }, { "epoch": 0.5338245144005358, "grad_norm": 2.4761008052076563, "learning_rate": 4.983102177743774e-06, "loss": 1.4962, "step": 797 }, { "epoch": 0.5344943067649028, "grad_norm": 2.2268818297333604, "learning_rate": 4.982932117671304e-06, "loss": 1.5325, "step": 798 }, { "epoch": 0.5351640991292699, "grad_norm": 2.221334869851946, "learning_rate": 4.982761209064823e-06, "loss": 1.8201, "step": 799 }, { "epoch": 0.535833891493637, "grad_norm": 2.8721716588767428, "learning_rate": 4.9825894519827364e-06, "loss": 1.8621, "step": 800 }, { "epoch": 0.5365036838580041, "grad_norm": 2.11264187235871, "learning_rate": 4.982416846483743e-06, "loss": 1.8027, "step": 801 }, { "epoch": 0.5371734762223711, "grad_norm": 3.2636642182002356, "learning_rate": 4.982243392626829e-06, "loss": 1.7914, "step": 802 }, { "epoch": 0.5378432685867381, "grad_norm": 4.893853241637807, "learning_rate": 4.982069090471273e-06, "loss": 1.8393, "step": 803 }, { "epoch": 0.5385130609511052, "grad_norm": 2.460965482968037, "learning_rate": 4.9818939400766405e-06, "loss": 1.6391, "step": 804 }, { "epoch": 0.5391828533154722, "grad_norm": 2.238262247618121, "learning_rate": 4.981717941502789e-06, "loss": 1.8404, "step": 805 }, { "epoch": 0.5398526456798393, "grad_norm": 2.832470854455245, "learning_rate": 4.9815410948098675e-06, "loss": 1.871, "step": 806 }, { "epoch": 0.5405224380442063, "grad_norm": 2.5138022009764995, "learning_rate": 4.98136340005831e-06, "loss": 1.5946, "step": 807 }, { "epoch": 0.5411922304085733, "grad_norm": 2.450488049000881, "learning_rate": 4.9811848573088454e-06, "loss": 2.1185, "step": 808 }, { "epoch": 0.5418620227729404, "grad_norm": 2.544087832031278, "learning_rate": 4.981005466622488e-06, "loss": 1.7382, "step": 809 }, { "epoch": 0.5425318151373074, "grad_norm": 2.9124078943448635, "learning_rate": 4.980825228060545e-06, "loss": 1.7686, "step": 810 }, { "epoch": 0.5432016075016745, "grad_norm": 2.8157405736647876, "learning_rate": 4.980644141684613e-06, "loss": 1.7945, "step": 811 }, { "epoch": 0.5438713998660415, "grad_norm": 2.187261464373384, "learning_rate": 4.9804622075565775e-06, "loss": 1.8566, "step": 812 }, { "epoch": 0.5445411922304085, "grad_norm": 2.1525730950653186, "learning_rate": 4.9802794257386125e-06, "loss": 1.7919, "step": 813 }, { "epoch": 0.5452109845947756, "grad_norm": 2.188244173107588, "learning_rate": 4.980095796293184e-06, "loss": 2.0093, "step": 814 }, { "epoch": 0.5458807769591426, "grad_norm": 1.9563949813563453, "learning_rate": 4.9799113192830475e-06, "loss": 1.6085, "step": 815 }, { "epoch": 0.5465505693235098, "grad_norm": 4.0518650059335695, "learning_rate": 4.979725994771246e-06, "loss": 1.7915, "step": 816 }, { "epoch": 0.5472203616878768, "grad_norm": 2.0362259196151955, "learning_rate": 4.979539822821115e-06, "loss": 1.7856, "step": 817 }, { "epoch": 0.5478901540522438, "grad_norm": 6.468138040707744, "learning_rate": 4.979352803496277e-06, "loss": 1.545, "step": 818 }, { "epoch": 0.5485599464166109, "grad_norm": 2.584149743862654, "learning_rate": 4.979164936860645e-06, "loss": 1.7845, "step": 819 }, { "epoch": 0.5492297387809779, "grad_norm": 3.5341081048892558, "learning_rate": 4.9789762229784235e-06, "loss": 1.6946, "step": 820 }, { "epoch": 0.549899531145345, "grad_norm": 3.8275375100505666, "learning_rate": 4.978786661914103e-06, "loss": 1.8005, "step": 821 }, { "epoch": 0.550569323509712, "grad_norm": 2.71402690018203, "learning_rate": 4.9785962537324674e-06, "loss": 1.6528, "step": 822 }, { "epoch": 0.551239115874079, "grad_norm": 2.237139464183015, "learning_rate": 4.978404998498586e-06, "loss": 1.9166, "step": 823 }, { "epoch": 0.5519089082384461, "grad_norm": 5.552342173274453, "learning_rate": 4.978212896277821e-06, "loss": 1.6238, "step": 824 }, { "epoch": 0.5525787006028131, "grad_norm": 2.187164273422503, "learning_rate": 4.978019947135823e-06, "loss": 1.766, "step": 825 }, { "epoch": 0.5532484929671801, "grad_norm": 2.2444124127559713, "learning_rate": 4.977826151138529e-06, "loss": 1.8013, "step": 826 }, { "epoch": 0.5539182853315472, "grad_norm": 2.3672053351932254, "learning_rate": 4.977631508352172e-06, "loss": 1.7974, "step": 827 }, { "epoch": 0.5545880776959142, "grad_norm": 2.1935867146266417, "learning_rate": 4.977436018843268e-06, "loss": 1.8109, "step": 828 }, { "epoch": 0.5552578700602813, "grad_norm": 2.086052501005329, "learning_rate": 4.977239682678626e-06, "loss": 1.8088, "step": 829 }, { "epoch": 0.5559276624246483, "grad_norm": 2.111989088361581, "learning_rate": 4.977042499925343e-06, "loss": 1.7796, "step": 830 }, { "epoch": 0.5565974547890155, "grad_norm": 2.285555033800255, "learning_rate": 4.976844470650804e-06, "loss": 1.7995, "step": 831 }, { "epoch": 0.5572672471533825, "grad_norm": 2.186332310923338, "learning_rate": 4.976645594922688e-06, "loss": 1.8587, "step": 832 }, { "epoch": 0.5579370395177495, "grad_norm": 2.181404352662693, "learning_rate": 4.976445872808958e-06, "loss": 1.811, "step": 833 }, { "epoch": 0.5586068318821166, "grad_norm": 3.4071691098343284, "learning_rate": 4.976245304377869e-06, "loss": 1.6768, "step": 834 }, { "epoch": 0.5592766242464836, "grad_norm": 4.458212466235543, "learning_rate": 4.976043889697964e-06, "loss": 1.8301, "step": 835 }, { "epoch": 0.5599464166108507, "grad_norm": 12.854983912479254, "learning_rate": 4.975841628838076e-06, "loss": 1.9563, "step": 836 }, { "epoch": 0.5606162089752177, "grad_norm": 3.2864914499481515, "learning_rate": 4.975638521867328e-06, "loss": 1.8313, "step": 837 }, { "epoch": 0.5612860013395847, "grad_norm": 2.2392029239104403, "learning_rate": 4.97543456885513e-06, "loss": 1.9241, "step": 838 }, { "epoch": 0.5619557937039518, "grad_norm": 2.1631842519729587, "learning_rate": 4.975229769871183e-06, "loss": 1.8363, "step": 839 }, { "epoch": 0.5626255860683188, "grad_norm": 2.1051206042839516, "learning_rate": 4.975024124985476e-06, "loss": 1.8, "step": 840 }, { "epoch": 0.5632953784326858, "grad_norm": 2.378409490733813, "learning_rate": 4.974817634268287e-06, "loss": 1.8641, "step": 841 }, { "epoch": 0.5639651707970529, "grad_norm": 2.2900148698760994, "learning_rate": 4.9746102977901845e-06, "loss": 1.6871, "step": 842 }, { "epoch": 0.5646349631614199, "grad_norm": 2.206913624670084, "learning_rate": 4.974402115622025e-06, "loss": 1.8739, "step": 843 }, { "epoch": 0.565304755525787, "grad_norm": 2.0735217605470204, "learning_rate": 4.974193087834953e-06, "loss": 1.6599, "step": 844 }, { "epoch": 0.565974547890154, "grad_norm": 2.3689424835032877, "learning_rate": 4.973983214500404e-06, "loss": 1.6238, "step": 845 }, { "epoch": 0.5666443402545212, "grad_norm": 3.7106239852775516, "learning_rate": 4.973772495690103e-06, "loss": 1.7775, "step": 846 }, { "epoch": 0.5673141326188882, "grad_norm": 3.7727481998742984, "learning_rate": 4.973560931476058e-06, "loss": 1.7985, "step": 847 }, { "epoch": 0.5679839249832552, "grad_norm": 2.6253744679091295, "learning_rate": 4.9733485219305746e-06, "loss": 1.2709, "step": 848 }, { "epoch": 0.5686537173476223, "grad_norm": 3.5240747720243477, "learning_rate": 4.973135267126243e-06, "loss": 1.6952, "step": 849 }, { "epoch": 0.5693235097119893, "grad_norm": 3.655036323892588, "learning_rate": 4.972921167135939e-06, "loss": 1.6746, "step": 850 }, { "epoch": 0.5699933020763563, "grad_norm": 6.237002804020971, "learning_rate": 4.9727062220328324e-06, "loss": 1.7113, "step": 851 }, { "epoch": 0.5706630944407234, "grad_norm": 2.6897669553321664, "learning_rate": 4.972490431890381e-06, "loss": 1.91, "step": 852 }, { "epoch": 0.5713328868050904, "grad_norm": 8.106909146690512, "learning_rate": 4.97227379678233e-06, "loss": 1.73, "step": 853 }, { "epoch": 0.5720026791694575, "grad_norm": 3.7335245762981386, "learning_rate": 4.972056316782713e-06, "loss": 1.8016, "step": 854 }, { "epoch": 0.5726724715338245, "grad_norm": 2.229225776856523, "learning_rate": 4.971837991965853e-06, "loss": 1.6663, "step": 855 }, { "epoch": 0.5733422638981915, "grad_norm": 6.759795434257313, "learning_rate": 4.971618822406363e-06, "loss": 1.9656, "step": 856 }, { "epoch": 0.5740120562625586, "grad_norm": 2.191508790218431, "learning_rate": 4.971398808179142e-06, "loss": 1.7539, "step": 857 }, { "epoch": 0.5746818486269256, "grad_norm": 2.432462024530815, "learning_rate": 4.9711779493593795e-06, "loss": 1.8182, "step": 858 }, { "epoch": 0.5753516409912927, "grad_norm": 3.8374954677638637, "learning_rate": 4.970956246022555e-06, "loss": 1.9593, "step": 859 }, { "epoch": 0.5760214333556597, "grad_norm": 3.4430238221501197, "learning_rate": 4.970733698244432e-06, "loss": 1.7727, "step": 860 }, { "epoch": 0.5766912257200268, "grad_norm": 2.246236834577686, "learning_rate": 4.970510306101067e-06, "loss": 1.8627, "step": 861 }, { "epoch": 0.5773610180843939, "grad_norm": 3.153047344458723, "learning_rate": 4.970286069668805e-06, "loss": 1.8291, "step": 862 }, { "epoch": 0.5780308104487609, "grad_norm": 2.2223113882838215, "learning_rate": 4.970060989024276e-06, "loss": 1.9395, "step": 863 }, { "epoch": 0.578700602813128, "grad_norm": 4.239884238998625, "learning_rate": 4.969835064244401e-06, "loss": 1.8198, "step": 864 }, { "epoch": 0.579370395177495, "grad_norm": 2.2713508147557206, "learning_rate": 4.969608295406389e-06, "loss": 1.8291, "step": 865 }, { "epoch": 0.580040187541862, "grad_norm": 3.928112665452493, "learning_rate": 4.969380682587738e-06, "loss": 1.6279, "step": 866 }, { "epoch": 0.5807099799062291, "grad_norm": 2.1502131584214195, "learning_rate": 4.969152225866234e-06, "loss": 1.7701, "step": 867 }, { "epoch": 0.5813797722705961, "grad_norm": 2.2065008636060806, "learning_rate": 4.968922925319951e-06, "loss": 1.6568, "step": 868 }, { "epoch": 0.5820495646349632, "grad_norm": 2.1043843820382597, "learning_rate": 4.968692781027251e-06, "loss": 1.8455, "step": 869 }, { "epoch": 0.5827193569993302, "grad_norm": 3.8849759530624763, "learning_rate": 4.968461793066787e-06, "loss": 1.637, "step": 870 }, { "epoch": 0.5833891493636972, "grad_norm": 2.3533969558200307, "learning_rate": 4.968229961517496e-06, "loss": 1.6363, "step": 871 }, { "epoch": 0.5840589417280643, "grad_norm": 3.1067544478180023, "learning_rate": 4.967997286458608e-06, "loss": 1.8649, "step": 872 }, { "epoch": 0.5847287340924313, "grad_norm": 3.96984512814224, "learning_rate": 4.967763767969638e-06, "loss": 1.757, "step": 873 }, { "epoch": 0.5853985264567984, "grad_norm": 2.198146540452862, "learning_rate": 4.967529406130389e-06, "loss": 1.6855, "step": 874 }, { "epoch": 0.5860683188211654, "grad_norm": 3.4100677056469095, "learning_rate": 4.967294201020954e-06, "loss": 1.8658, "step": 875 }, { "epoch": 0.5867381111855325, "grad_norm": 2.0339965158795144, "learning_rate": 4.967058152721716e-06, "loss": 1.6853, "step": 876 }, { "epoch": 0.5874079035498996, "grad_norm": 9.986021388198798, "learning_rate": 4.966821261313341e-06, "loss": 1.7107, "step": 877 }, { "epoch": 0.5880776959142666, "grad_norm": 2.256593466530863, "learning_rate": 4.966583526876786e-06, "loss": 1.6497, "step": 878 }, { "epoch": 0.5887474882786337, "grad_norm": 4.3456106189132155, "learning_rate": 4.966344949493297e-06, "loss": 1.7488, "step": 879 }, { "epoch": 0.5894172806430007, "grad_norm": 3.186830275867281, "learning_rate": 4.966105529244407e-06, "loss": 1.8609, "step": 880 }, { "epoch": 0.5900870730073677, "grad_norm": 2.4101253784303096, "learning_rate": 4.965865266211936e-06, "loss": 1.5376, "step": 881 }, { "epoch": 0.5907568653717348, "grad_norm": 2.6926830417022387, "learning_rate": 4.965624160477995e-06, "loss": 1.6825, "step": 882 }, { "epoch": 0.5914266577361018, "grad_norm": 2.3613200605257942, "learning_rate": 4.96538221212498e-06, "loss": 1.6849, "step": 883 }, { "epoch": 0.5920964501004689, "grad_norm": 2.2873676950180224, "learning_rate": 4.965139421235575e-06, "loss": 1.7703, "step": 884 }, { "epoch": 0.5927662424648359, "grad_norm": 2.2060391227727534, "learning_rate": 4.964895787892755e-06, "loss": 1.7338, "step": 885 }, { "epoch": 0.5934360348292029, "grad_norm": 3.161204159960425, "learning_rate": 4.964651312179779e-06, "loss": 1.9226, "step": 886 }, { "epoch": 0.59410582719357, "grad_norm": 4.639579205387801, "learning_rate": 4.964405994180197e-06, "loss": 1.6267, "step": 887 }, { "epoch": 0.594775619557937, "grad_norm": 2.8831402100347283, "learning_rate": 4.964159833977846e-06, "loss": 1.6513, "step": 888 }, { "epoch": 0.595445411922304, "grad_norm": 2.6712087130595554, "learning_rate": 4.96391283165685e-06, "loss": 1.8709, "step": 889 }, { "epoch": 0.5961152042866711, "grad_norm": 2.2733431261975796, "learning_rate": 4.96366498730162e-06, "loss": 1.6421, "step": 890 }, { "epoch": 0.5967849966510381, "grad_norm": 2.2251329847866574, "learning_rate": 4.963416300996857e-06, "loss": 1.7896, "step": 891 }, { "epoch": 0.5974547890154053, "grad_norm": 5.3367101637672985, "learning_rate": 4.96316677282755e-06, "loss": 1.9046, "step": 892 }, { "epoch": 0.5981245813797723, "grad_norm": 2.2973620055062236, "learning_rate": 4.962916402878972e-06, "loss": 1.7982, "step": 893 }, { "epoch": 0.5987943737441394, "grad_norm": 2.617855217797549, "learning_rate": 4.9626651912366885e-06, "loss": 1.7916, "step": 894 }, { "epoch": 0.5994641661085064, "grad_norm": 3.224787019685552, "learning_rate": 4.962413137986549e-06, "loss": 1.5524, "step": 895 }, { "epoch": 0.6001339584728734, "grad_norm": 2.3864499322626953, "learning_rate": 4.962160243214692e-06, "loss": 1.7512, "step": 896 }, { "epoch": 0.6008037508372405, "grad_norm": 3.4972877814754013, "learning_rate": 4.9619065070075435e-06, "loss": 1.7395, "step": 897 }, { "epoch": 0.6014735432016075, "grad_norm": 2.939358488445085, "learning_rate": 4.9616519294518166e-06, "loss": 1.802, "step": 898 }, { "epoch": 0.6021433355659745, "grad_norm": 3.39472790673276, "learning_rate": 4.961396510634513e-06, "loss": 1.7668, "step": 899 }, { "epoch": 0.6028131279303416, "grad_norm": 2.4167500983051737, "learning_rate": 4.961140250642921e-06, "loss": 1.7521, "step": 900 }, { "epoch": 0.6034829202947086, "grad_norm": 3.1119547615078686, "learning_rate": 4.960883149564616e-06, "loss": 1.6709, "step": 901 }, { "epoch": 0.6041527126590757, "grad_norm": 3.9177914171343815, "learning_rate": 4.9606252074874624e-06, "loss": 1.5821, "step": 902 }, { "epoch": 0.6048225050234427, "grad_norm": 2.1363695134324656, "learning_rate": 4.9603664244996115e-06, "loss": 1.8477, "step": 903 }, { "epoch": 0.6054922973878097, "grad_norm": 2.5230837613323103, "learning_rate": 4.960106800689501e-06, "loss": 1.6947, "step": 904 }, { "epoch": 0.6061620897521768, "grad_norm": 2.271363014238947, "learning_rate": 4.9598463361458556e-06, "loss": 1.7612, "step": 905 }, { "epoch": 0.6068318821165438, "grad_norm": 4.155663320914737, "learning_rate": 4.959585030957689e-06, "loss": 1.7675, "step": 906 }, { "epoch": 0.607501674480911, "grad_norm": 2.4029321607563077, "learning_rate": 4.959322885214302e-06, "loss": 1.6517, "step": 907 }, { "epoch": 0.608171466845278, "grad_norm": 3.330719073310976, "learning_rate": 4.95905989900528e-06, "loss": 1.8004, "step": 908 }, { "epoch": 0.608841259209645, "grad_norm": 2.6495772673218476, "learning_rate": 4.958796072420501e-06, "loss": 1.5149, "step": 909 }, { "epoch": 0.6095110515740121, "grad_norm": 2.25402286220689, "learning_rate": 4.958531405550124e-06, "loss": 1.6314, "step": 910 }, { "epoch": 0.6101808439383791, "grad_norm": 2.0994861536162963, "learning_rate": 4.958265898484599e-06, "loss": 1.5717, "step": 911 }, { "epoch": 0.6108506363027462, "grad_norm": 2.7521991308260216, "learning_rate": 4.957999551314662e-06, "loss": 1.5775, "step": 912 }, { "epoch": 0.6115204286671132, "grad_norm": 2.167169818179813, "learning_rate": 4.957732364131337e-06, "loss": 1.6408, "step": 913 }, { "epoch": 0.6121902210314802, "grad_norm": 1.9826314246010046, "learning_rate": 4.957464337025934e-06, "loss": 1.609, "step": 914 }, { "epoch": 0.6128600133958473, "grad_norm": 5.332736616518463, "learning_rate": 4.957195470090049e-06, "loss": 1.8617, "step": 915 }, { "epoch": 0.6135298057602143, "grad_norm": 2.1284141507974463, "learning_rate": 4.956925763415569e-06, "loss": 1.7606, "step": 916 }, { "epoch": 0.6141995981245814, "grad_norm": 2.5107737462602615, "learning_rate": 4.956655217094663e-06, "loss": 1.7742, "step": 917 }, { "epoch": 0.6148693904889484, "grad_norm": 6.44205156052387, "learning_rate": 4.956383831219791e-06, "loss": 1.6273, "step": 918 }, { "epoch": 0.6155391828533154, "grad_norm": 4.967949981828972, "learning_rate": 4.956111605883697e-06, "loss": 1.8667, "step": 919 }, { "epoch": 0.6162089752176825, "grad_norm": 4.302926203916212, "learning_rate": 4.955838541179414e-06, "loss": 1.8472, "step": 920 }, { "epoch": 0.6168787675820495, "grad_norm": 3.850992071936025, "learning_rate": 4.955564637200261e-06, "loss": 1.6441, "step": 921 }, { "epoch": 0.6175485599464167, "grad_norm": 2.261050202374149, "learning_rate": 4.955289894039843e-06, "loss": 1.7191, "step": 922 }, { "epoch": 0.6182183523107837, "grad_norm": 4.699032148243999, "learning_rate": 4.955014311792054e-06, "loss": 1.7381, "step": 923 }, { "epoch": 0.6188881446751507, "grad_norm": 3.1606642982765254, "learning_rate": 4.954737890551072e-06, "loss": 1.7373, "step": 924 }, { "epoch": 0.6195579370395178, "grad_norm": 30.951546560399468, "learning_rate": 4.954460630411363e-06, "loss": 1.7393, "step": 925 }, { "epoch": 0.6202277294038848, "grad_norm": 3.3416767156698315, "learning_rate": 4.954182531467681e-06, "loss": 1.6466, "step": 926 }, { "epoch": 0.6208975217682519, "grad_norm": 2.9882197777415915, "learning_rate": 4.953903593815063e-06, "loss": 1.8069, "step": 927 }, { "epoch": 0.6215673141326189, "grad_norm": 2.4821437768310677, "learning_rate": 4.953623817548839e-06, "loss": 1.8863, "step": 928 }, { "epoch": 0.6222371064969859, "grad_norm": 2.087707846257214, "learning_rate": 4.953343202764618e-06, "loss": 1.9113, "step": 929 }, { "epoch": 0.622906898861353, "grad_norm": 2.7704940008697525, "learning_rate": 4.953061749558301e-06, "loss": 1.6024, "step": 930 }, { "epoch": 0.62357669122572, "grad_norm": 2.1548966632273707, "learning_rate": 4.952779458026073e-06, "loss": 1.8595, "step": 931 }, { "epoch": 0.624246483590087, "grad_norm": 2.46166377814114, "learning_rate": 4.952496328264407e-06, "loss": 1.8223, "step": 932 }, { "epoch": 0.6249162759544541, "grad_norm": 2.4029359723066364, "learning_rate": 4.95221236037006e-06, "loss": 1.7262, "step": 933 }, { "epoch": 0.6255860683188211, "grad_norm": 2.2416618653704106, "learning_rate": 4.951927554440079e-06, "loss": 1.6374, "step": 934 }, { "epoch": 0.6262558606831882, "grad_norm": 2.8863802831918686, "learning_rate": 4.951641910571795e-06, "loss": 1.7218, "step": 935 }, { "epoch": 0.6269256530475552, "grad_norm": 3.0072717526952584, "learning_rate": 4.951355428862825e-06, "loss": 1.7636, "step": 936 }, { "epoch": 0.6275954454119224, "grad_norm": 4.15295480217641, "learning_rate": 4.9510681094110746e-06, "loss": 1.6135, "step": 937 }, { "epoch": 0.6282652377762894, "grad_norm": 3.780609764532968, "learning_rate": 4.950779952314732e-06, "loss": 1.7753, "step": 938 }, { "epoch": 0.6289350301406564, "grad_norm": 2.261633331068276, "learning_rate": 4.950490957672277e-06, "loss": 1.8191, "step": 939 }, { "epoch": 0.6296048225050235, "grad_norm": 3.228882325548011, "learning_rate": 4.950201125582471e-06, "loss": 1.8521, "step": 940 }, { "epoch": 0.6302746148693905, "grad_norm": 5.092329029162018, "learning_rate": 4.949910456144362e-06, "loss": 1.9015, "step": 941 }, { "epoch": 0.6309444072337576, "grad_norm": 3.183492015428902, "learning_rate": 4.949618949457288e-06, "loss": 1.6576, "step": 942 }, { "epoch": 0.6316141995981246, "grad_norm": 2.408898230028661, "learning_rate": 4.949326605620868e-06, "loss": 1.4736, "step": 943 }, { "epoch": 0.6322839919624916, "grad_norm": 5.592853874115134, "learning_rate": 4.949033424735012e-06, "loss": 1.6483, "step": 944 }, { "epoch": 0.6329537843268587, "grad_norm": 2.5090832156979603, "learning_rate": 4.94873940689991e-06, "loss": 1.5459, "step": 945 }, { "epoch": 0.6336235766912257, "grad_norm": 2.609476356139649, "learning_rate": 4.948444552216045e-06, "loss": 1.6279, "step": 946 }, { "epoch": 0.6342933690555927, "grad_norm": 2.182110607061972, "learning_rate": 4.948148860784182e-06, "loss": 1.6302, "step": 947 }, { "epoch": 0.6349631614199598, "grad_norm": 2.1345836711078494, "learning_rate": 4.947852332705372e-06, "loss": 1.5353, "step": 948 }, { "epoch": 0.6356329537843268, "grad_norm": 2.0784594059025183, "learning_rate": 4.947554968080952e-06, "loss": 1.8619, "step": 949 }, { "epoch": 0.6363027461486939, "grad_norm": 3.714787107506573, "learning_rate": 4.947256767012546e-06, "loss": 1.7341, "step": 950 }, { "epoch": 0.6369725385130609, "grad_norm": 2.808968112134475, "learning_rate": 4.946957729602063e-06, "loss": 1.7212, "step": 951 }, { "epoch": 0.6376423308774279, "grad_norm": 8.789865925057944, "learning_rate": 4.946657855951699e-06, "loss": 1.8331, "step": 952 }, { "epoch": 0.6383121232417951, "grad_norm": 3.298263836493085, "learning_rate": 4.946357146163934e-06, "loss": 1.6143, "step": 953 }, { "epoch": 0.6389819156061621, "grad_norm": 2.4015399559179085, "learning_rate": 4.946055600341535e-06, "loss": 1.7606, "step": 954 }, { "epoch": 0.6396517079705292, "grad_norm": 2.6590931888600435, "learning_rate": 4.945753218587553e-06, "loss": 1.8444, "step": 955 }, { "epoch": 0.6403215003348962, "grad_norm": 2.131998145270563, "learning_rate": 4.945450001005328e-06, "loss": 1.8168, "step": 956 }, { "epoch": 0.6409912926992632, "grad_norm": 3.108715985243645, "learning_rate": 4.945145947698483e-06, "loss": 1.7804, "step": 957 }, { "epoch": 0.6416610850636303, "grad_norm": 2.470466445793264, "learning_rate": 4.944841058770926e-06, "loss": 1.7856, "step": 958 }, { "epoch": 0.6423308774279973, "grad_norm": 3.5400124957836754, "learning_rate": 4.944535334326853e-06, "loss": 1.8618, "step": 959 }, { "epoch": 0.6430006697923644, "grad_norm": 2.099083079850069, "learning_rate": 4.944228774470744e-06, "loss": 1.7171, "step": 960 }, { "epoch": 0.6436704621567314, "grad_norm": 2.1355983967052, "learning_rate": 4.943921379307365e-06, "loss": 1.6923, "step": 961 }, { "epoch": 0.6443402545210984, "grad_norm": 2.1072422516809493, "learning_rate": 4.943613148941768e-06, "loss": 1.764, "step": 962 }, { "epoch": 0.6450100468854655, "grad_norm": 3.171259365265046, "learning_rate": 4.943304083479288e-06, "loss": 1.7462, "step": 963 }, { "epoch": 0.6456798392498325, "grad_norm": 4.549106940088939, "learning_rate": 4.942994183025548e-06, "loss": 1.8149, "step": 964 }, { "epoch": 0.6463496316141996, "grad_norm": 2.371573335568395, "learning_rate": 4.942683447686456e-06, "loss": 1.8286, "step": 965 }, { "epoch": 0.6470194239785666, "grad_norm": 2.5573647306743936, "learning_rate": 4.942371877568203e-06, "loss": 1.8957, "step": 966 }, { "epoch": 0.6476892163429336, "grad_norm": 2.896059871898815, "learning_rate": 4.94205947277727e-06, "loss": 1.6441, "step": 967 }, { "epoch": 0.6483590087073008, "grad_norm": 7.107335284644982, "learning_rate": 4.941746233420418e-06, "loss": 1.7757, "step": 968 }, { "epoch": 0.6490288010716678, "grad_norm": 4.787905946785729, "learning_rate": 4.941432159604697e-06, "loss": 1.6113, "step": 969 }, { "epoch": 0.6496985934360349, "grad_norm": 2.440139733981447, "learning_rate": 4.941117251437438e-06, "loss": 1.7395, "step": 970 }, { "epoch": 0.6503683858004019, "grad_norm": 2.114980142860102, "learning_rate": 4.940801509026264e-06, "loss": 1.7414, "step": 971 }, { "epoch": 0.6510381781647689, "grad_norm": 3.0877203255937657, "learning_rate": 4.940484932479075e-06, "loss": 1.7466, "step": 972 }, { "epoch": 0.651707970529136, "grad_norm": 2.199922429866672, "learning_rate": 4.940167521904064e-06, "loss": 1.5211, "step": 973 }, { "epoch": 0.652377762893503, "grad_norm": 3.537526929988689, "learning_rate": 4.939849277409701e-06, "loss": 1.7158, "step": 974 }, { "epoch": 0.6530475552578701, "grad_norm": 2.5626181660789995, "learning_rate": 4.9395301991047465e-06, "loss": 1.5712, "step": 975 }, { "epoch": 0.6537173476222371, "grad_norm": 2.46272367840424, "learning_rate": 4.939210287098246e-06, "loss": 2.0372, "step": 976 }, { "epoch": 0.6543871399866041, "grad_norm": 2.6210488598543304, "learning_rate": 4.938889541499526e-06, "loss": 1.8627, "step": 977 }, { "epoch": 0.6550569323509712, "grad_norm": 2.192107244751343, "learning_rate": 4.938567962418202e-06, "loss": 1.8854, "step": 978 }, { "epoch": 0.6557267247153382, "grad_norm": 2.162874848111986, "learning_rate": 4.938245549964174e-06, "loss": 1.7689, "step": 979 }, { "epoch": 0.6563965170797053, "grad_norm": 2.139672896080246, "learning_rate": 4.937922304247621e-06, "loss": 1.7415, "step": 980 }, { "epoch": 0.6570663094440723, "grad_norm": 2.925033705245606, "learning_rate": 4.937598225379015e-06, "loss": 1.7752, "step": 981 }, { "epoch": 0.6577361018084393, "grad_norm": 2.8432838406250314, "learning_rate": 4.9372733134691065e-06, "loss": 1.7094, "step": 982 }, { "epoch": 0.6584058941728065, "grad_norm": 7.84329965393524, "learning_rate": 4.9369475686289345e-06, "loss": 1.6924, "step": 983 }, { "epoch": 0.6590756865371735, "grad_norm": 2.735656566273711, "learning_rate": 4.936620990969822e-06, "loss": 1.6373, "step": 984 }, { "epoch": 0.6597454789015406, "grad_norm": 2.7217817028476903, "learning_rate": 4.9362935806033726e-06, "loss": 1.8152, "step": 985 }, { "epoch": 0.6604152712659076, "grad_norm": 2.1837900975563267, "learning_rate": 4.9359653376414815e-06, "loss": 1.7563, "step": 986 }, { "epoch": 0.6610850636302746, "grad_norm": 2.2807172272771954, "learning_rate": 4.935636262196323e-06, "loss": 1.5701, "step": 987 }, { "epoch": 0.6617548559946417, "grad_norm": 3.6921601400783497, "learning_rate": 4.935306354380357e-06, "loss": 1.4518, "step": 988 }, { "epoch": 0.6624246483590087, "grad_norm": 2.9367603377581215, "learning_rate": 4.9349756143063295e-06, "loss": 1.8319, "step": 989 }, { "epoch": 0.6630944407233758, "grad_norm": 2.5416935513013286, "learning_rate": 4.93464404208727e-06, "loss": 1.6413, "step": 990 }, { "epoch": 0.6637642330877428, "grad_norm": 2.664411184039861, "learning_rate": 4.934311637836491e-06, "loss": 1.7485, "step": 991 }, { "epoch": 0.6644340254521098, "grad_norm": 3.5323928912089544, "learning_rate": 4.93397840166759e-06, "loss": 1.7626, "step": 992 }, { "epoch": 0.6651038178164769, "grad_norm": 2.19016306326781, "learning_rate": 4.933644333694453e-06, "loss": 1.8423, "step": 993 }, { "epoch": 0.6657736101808439, "grad_norm": 2.190302761520974, "learning_rate": 4.933309434031243e-06, "loss": 1.8337, "step": 994 }, { "epoch": 0.666443402545211, "grad_norm": 18.066341733671713, "learning_rate": 4.932973702792412e-06, "loss": 1.7557, "step": 995 }, { "epoch": 0.667113194909578, "grad_norm": 4.653363677902187, "learning_rate": 4.9326371400926955e-06, "loss": 1.7542, "step": 996 }, { "epoch": 0.667782987273945, "grad_norm": 3.36874751688877, "learning_rate": 4.932299746047112e-06, "loss": 1.7237, "step": 997 }, { "epoch": 0.6684527796383122, "grad_norm": 2.3010280504178096, "learning_rate": 4.931961520770966e-06, "loss": 1.7385, "step": 998 }, { "epoch": 0.6691225720026792, "grad_norm": 2.28998357369975, "learning_rate": 4.931622464379843e-06, "loss": 1.6023, "step": 999 }, { "epoch": 0.6697923643670463, "grad_norm": 3.5624276176001635, "learning_rate": 4.9312825769896166e-06, "loss": 1.5092, "step": 1000 }, { "epoch": 0.6704621567314133, "grad_norm": 2.295901381443903, "learning_rate": 4.9309418587164405e-06, "loss": 1.8987, "step": 1001 }, { "epoch": 0.6711319490957803, "grad_norm": 3.5605281011290937, "learning_rate": 4.930600309676755e-06, "loss": 1.7558, "step": 1002 }, { "epoch": 0.6718017414601474, "grad_norm": 5.582037308527117, "learning_rate": 4.930257929987283e-06, "loss": 1.7867, "step": 1003 }, { "epoch": 0.6724715338245144, "grad_norm": 2.1383644999345375, "learning_rate": 4.929914719765032e-06, "loss": 1.814, "step": 1004 }, { "epoch": 0.6731413261888815, "grad_norm": 16.307032861938584, "learning_rate": 4.929570679127292e-06, "loss": 1.581, "step": 1005 }, { "epoch": 0.6738111185532485, "grad_norm": 3.4328599232096537, "learning_rate": 4.92922580819164e-06, "loss": 1.6102, "step": 1006 }, { "epoch": 0.6744809109176155, "grad_norm": 3.70635903418262, "learning_rate": 4.928880107075932e-06, "loss": 1.598, "step": 1007 }, { "epoch": 0.6751507032819826, "grad_norm": 4.392842784379396, "learning_rate": 4.928533575898311e-06, "loss": 1.615, "step": 1008 }, { "epoch": 0.6758204956463496, "grad_norm": 6.384031655311987, "learning_rate": 4.928186214777204e-06, "loss": 1.7588, "step": 1009 }, { "epoch": 0.6764902880107166, "grad_norm": 3.649398523080596, "learning_rate": 4.92783802383132e-06, "loss": 1.7336, "step": 1010 }, { "epoch": 0.6771600803750837, "grad_norm": 2.320179660234767, "learning_rate": 4.9274890031796505e-06, "loss": 1.8427, "step": 1011 }, { "epoch": 0.6778298727394507, "grad_norm": 2.2109426139037835, "learning_rate": 4.927139152941474e-06, "loss": 1.8663, "step": 1012 }, { "epoch": 0.6784996651038178, "grad_norm": 2.0196325536018414, "learning_rate": 4.92678847323635e-06, "loss": 1.7822, "step": 1013 }, { "epoch": 0.6791694574681849, "grad_norm": 2.357615436869285, "learning_rate": 4.926436964184124e-06, "loss": 1.5572, "step": 1014 }, { "epoch": 0.679839249832552, "grad_norm": 2.414525958206758, "learning_rate": 4.92608462590492e-06, "loss": 1.7963, "step": 1015 }, { "epoch": 0.680509042196919, "grad_norm": 4.080140962418607, "learning_rate": 4.9257314585191496e-06, "loss": 1.7541, "step": 1016 }, { "epoch": 0.681178834561286, "grad_norm": 12.104756021916621, "learning_rate": 4.9253774621475075e-06, "loss": 1.641, "step": 1017 }, { "epoch": 0.6818486269256531, "grad_norm": 4.633261666656377, "learning_rate": 4.92502263691097e-06, "loss": 1.5579, "step": 1018 }, { "epoch": 0.6825184192900201, "grad_norm": 2.065825635205898, "learning_rate": 4.924666982930798e-06, "loss": 1.8943, "step": 1019 }, { "epoch": 0.6831882116543871, "grad_norm": 4.642269031338428, "learning_rate": 4.924310500328535e-06, "loss": 1.8702, "step": 1020 }, { "epoch": 0.6838580040187542, "grad_norm": 3.2238569553273027, "learning_rate": 4.923953189226007e-06, "loss": 1.7369, "step": 1021 }, { "epoch": 0.6845277963831212, "grad_norm": 4.293472438940969, "learning_rate": 4.923595049745325e-06, "loss": 1.8625, "step": 1022 }, { "epoch": 0.6851975887474883, "grad_norm": 3.812115747234485, "learning_rate": 4.923236082008881e-06, "loss": 1.7755, "step": 1023 }, { "epoch": 0.6858673811118553, "grad_norm": 5.747607837073481, "learning_rate": 4.922876286139353e-06, "loss": 1.5919, "step": 1024 }, { "epoch": 0.6865371734762223, "grad_norm": 2.2132460036117507, "learning_rate": 4.922515662259697e-06, "loss": 1.6298, "step": 1025 }, { "epoch": 0.6872069658405894, "grad_norm": 3.271754368262585, "learning_rate": 4.922154210493158e-06, "loss": 1.7394, "step": 1026 }, { "epoch": 0.6878767582049564, "grad_norm": 5.963774225896432, "learning_rate": 4.921791930963258e-06, "loss": 1.7318, "step": 1027 }, { "epoch": 0.6885465505693235, "grad_norm": 3.4461162571266164, "learning_rate": 4.921428823793807e-06, "loss": 1.761, "step": 1028 }, { "epoch": 0.6892163429336906, "grad_norm": 2.9983843735260947, "learning_rate": 4.9210648891088965e-06, "loss": 1.4679, "step": 1029 }, { "epoch": 0.6898861352980576, "grad_norm": 4.560400009984782, "learning_rate": 4.920700127032897e-06, "loss": 1.7799, "step": 1030 }, { "epoch": 0.6905559276624247, "grad_norm": 4.052211507249368, "learning_rate": 4.920334537690468e-06, "loss": 1.6188, "step": 1031 }, { "epoch": 0.6912257200267917, "grad_norm": 9.27187117920403, "learning_rate": 4.919968121206546e-06, "loss": 1.7834, "step": 1032 }, { "epoch": 0.6918955123911588, "grad_norm": 2.3466125383479226, "learning_rate": 4.9196008777063535e-06, "loss": 1.3886, "step": 1033 }, { "epoch": 0.6925653047555258, "grad_norm": 2.7763534687525917, "learning_rate": 4.919232807315396e-06, "loss": 1.8543, "step": 1034 }, { "epoch": 0.6932350971198928, "grad_norm": 2.3377686323187854, "learning_rate": 4.918863910159458e-06, "loss": 1.7255, "step": 1035 }, { "epoch": 0.6939048894842599, "grad_norm": 3.352761508634192, "learning_rate": 4.918494186364611e-06, "loss": 1.8391, "step": 1036 }, { "epoch": 0.6945746818486269, "grad_norm": 2.2283636606480743, "learning_rate": 4.9181236360572045e-06, "loss": 1.7464, "step": 1037 }, { "epoch": 0.695244474212994, "grad_norm": 2.604765320946762, "learning_rate": 4.917752259363876e-06, "loss": 1.7229, "step": 1038 }, { "epoch": 0.695914266577361, "grad_norm": 3.247874994406768, "learning_rate": 4.9173800564115405e-06, "loss": 1.756, "step": 1039 }, { "epoch": 0.696584058941728, "grad_norm": 2.1317832765545974, "learning_rate": 4.917007027327398e-06, "loss": 1.8689, "step": 1040 }, { "epoch": 0.6972538513060951, "grad_norm": 5.547538957504823, "learning_rate": 4.916633172238927e-06, "loss": 1.7396, "step": 1041 }, { "epoch": 0.6979236436704621, "grad_norm": 6.565730778541284, "learning_rate": 4.916258491273895e-06, "loss": 1.6258, "step": 1042 }, { "epoch": 0.6985934360348292, "grad_norm": 3.3019997003491754, "learning_rate": 4.915882984560347e-06, "loss": 1.6075, "step": 1043 }, { "epoch": 0.6992632283991963, "grad_norm": 2.606212331658262, "learning_rate": 4.915506652226611e-06, "loss": 1.6216, "step": 1044 }, { "epoch": 0.6999330207635633, "grad_norm": 2.839010387119171, "learning_rate": 4.9151294944012965e-06, "loss": 1.7004, "step": 1045 }, { "epoch": 0.7006028131279304, "grad_norm": 2.469250466756659, "learning_rate": 4.914751511213296e-06, "loss": 1.6734, "step": 1046 }, { "epoch": 0.7012726054922974, "grad_norm": 3.482858848449481, "learning_rate": 4.914372702791786e-06, "loss": 1.8647, "step": 1047 }, { "epoch": 0.7019423978566645, "grad_norm": 3.6035379755920385, "learning_rate": 4.913993069266221e-06, "loss": 1.6778, "step": 1048 }, { "epoch": 0.7026121902210315, "grad_norm": 2.3282571844252953, "learning_rate": 4.913612610766341e-06, "loss": 1.8282, "step": 1049 }, { "epoch": 0.7032819825853985, "grad_norm": 4.595463997092846, "learning_rate": 4.913231327422164e-06, "loss": 1.6668, "step": 1050 }, { "epoch": 0.7039517749497656, "grad_norm": 4.615889492626988, "learning_rate": 4.912849219363996e-06, "loss": 1.7071, "step": 1051 }, { "epoch": 0.7046215673141326, "grad_norm": 2.171270780695349, "learning_rate": 4.912466286722418e-06, "loss": 1.8952, "step": 1052 }, { "epoch": 0.7052913596784997, "grad_norm": 2.4335521554407413, "learning_rate": 4.912082529628298e-06, "loss": 1.7399, "step": 1053 }, { "epoch": 0.7059611520428667, "grad_norm": 2.1689784135567103, "learning_rate": 4.911697948212783e-06, "loss": 1.6704, "step": 1054 }, { "epoch": 0.7066309444072337, "grad_norm": 3.6224046490749546, "learning_rate": 4.911312542607301e-06, "loss": 1.5974, "step": 1055 }, { "epoch": 0.7073007367716008, "grad_norm": 6.320381868685609, "learning_rate": 4.910926312943566e-06, "loss": 1.6855, "step": 1056 }, { "epoch": 0.7079705291359678, "grad_norm": 2.2697354789830593, "learning_rate": 4.910539259353569e-06, "loss": 1.648, "step": 1057 }, { "epoch": 0.7086403215003348, "grad_norm": 2.231496440672171, "learning_rate": 4.910151381969585e-06, "loss": 1.6316, "step": 1058 }, { "epoch": 0.709310113864702, "grad_norm": 2.164917330522565, "learning_rate": 4.909762680924168e-06, "loss": 1.7126, "step": 1059 }, { "epoch": 0.709979906229069, "grad_norm": 2.2182698980029274, "learning_rate": 4.909373156350159e-06, "loss": 1.6863, "step": 1060 }, { "epoch": 0.7106496985934361, "grad_norm": 2.06061409461232, "learning_rate": 4.908982808380672e-06, "loss": 1.6964, "step": 1061 }, { "epoch": 0.7113194909578031, "grad_norm": 4.550090481438596, "learning_rate": 4.908591637149112e-06, "loss": 1.8424, "step": 1062 }, { "epoch": 0.7119892833221702, "grad_norm": 13.366785454610289, "learning_rate": 4.908199642789156e-06, "loss": 1.5605, "step": 1063 }, { "epoch": 0.7126590756865372, "grad_norm": 2.336722451636987, "learning_rate": 4.907806825434771e-06, "loss": 1.8273, "step": 1064 }, { "epoch": 0.7133288680509042, "grad_norm": 2.870910426188679, "learning_rate": 4.907413185220198e-06, "loss": 1.6497, "step": 1065 }, { "epoch": 0.7139986604152713, "grad_norm": 2.2003816960331104, "learning_rate": 4.907018722279964e-06, "loss": 1.7459, "step": 1066 }, { "epoch": 0.7146684527796383, "grad_norm": 2.2812368702637107, "learning_rate": 4.906623436748874e-06, "loss": 1.7393, "step": 1067 }, { "epoch": 0.7153382451440053, "grad_norm": 4.191125152575635, "learning_rate": 4.906227328762017e-06, "loss": 1.695, "step": 1068 }, { "epoch": 0.7160080375083724, "grad_norm": 2.521008601388981, "learning_rate": 4.90583039845476e-06, "loss": 1.7542, "step": 1069 }, { "epoch": 0.7166778298727394, "grad_norm": 2.7576735715781795, "learning_rate": 4.905432645962754e-06, "loss": 1.6419, "step": 1070 }, { "epoch": 0.7173476222371065, "grad_norm": 3.4850183676808086, "learning_rate": 4.905034071421929e-06, "loss": 1.5345, "step": 1071 }, { "epoch": 0.7180174146014735, "grad_norm": 4.126250402790044, "learning_rate": 4.904634674968497e-06, "loss": 1.8125, "step": 1072 }, { "epoch": 0.7186872069658405, "grad_norm": 2.7523089235171327, "learning_rate": 4.904234456738951e-06, "loss": 1.5726, "step": 1073 }, { "epoch": 0.7193569993302076, "grad_norm": 2.304346788950656, "learning_rate": 4.903833416870062e-06, "loss": 1.6331, "step": 1074 }, { "epoch": 0.7200267916945747, "grad_norm": 2.1783507400039315, "learning_rate": 4.903431555498887e-06, "loss": 1.6843, "step": 1075 }, { "epoch": 0.7206965840589418, "grad_norm": 4.389552073456667, "learning_rate": 4.903028872762758e-06, "loss": 1.9899, "step": 1076 }, { "epoch": 0.7213663764233088, "grad_norm": 2.085672577486805, "learning_rate": 4.9026253687992925e-06, "loss": 1.7018, "step": 1077 }, { "epoch": 0.7220361687876758, "grad_norm": 6.452616207530919, "learning_rate": 4.9022210437463856e-06, "loss": 1.9349, "step": 1078 }, { "epoch": 0.7227059611520429, "grad_norm": 2.3662879107392616, "learning_rate": 4.901815897742216e-06, "loss": 1.6256, "step": 1079 }, { "epoch": 0.7233757535164099, "grad_norm": 3.025468924965611, "learning_rate": 4.901409930925238e-06, "loss": 1.7635, "step": 1080 }, { "epoch": 0.724045545880777, "grad_norm": 2.2931247765307607, "learning_rate": 4.901003143434192e-06, "loss": 1.7686, "step": 1081 }, { "epoch": 0.724715338245144, "grad_norm": 2.2513152536427135, "learning_rate": 4.900595535408095e-06, "loss": 1.7855, "step": 1082 }, { "epoch": 0.725385130609511, "grad_norm": 3.2166074385788783, "learning_rate": 4.900187106986246e-06, "loss": 1.6156, "step": 1083 }, { "epoch": 0.7260549229738781, "grad_norm": 2.6255663216143392, "learning_rate": 4.899777858308224e-06, "loss": 1.9043, "step": 1084 }, { "epoch": 0.7267247153382451, "grad_norm": 2.886666339637203, "learning_rate": 4.899367789513889e-06, "loss": 1.5491, "step": 1085 }, { "epoch": 0.7273945077026122, "grad_norm": 4.891410624938784, "learning_rate": 4.89895690074338e-06, "loss": 1.6784, "step": 1086 }, { "epoch": 0.7280643000669792, "grad_norm": 3.1552836328000544, "learning_rate": 4.898545192137117e-06, "loss": 1.6978, "step": 1087 }, { "epoch": 0.7287340924313462, "grad_norm": 3.341219654504165, "learning_rate": 4.898132663835801e-06, "loss": 1.6751, "step": 1088 }, { "epoch": 0.7294038847957133, "grad_norm": 2.4532150210060792, "learning_rate": 4.8977193159804106e-06, "loss": 1.8687, "step": 1089 }, { "epoch": 0.7300736771600804, "grad_norm": 2.2502811454079517, "learning_rate": 4.897305148712207e-06, "loss": 1.7601, "step": 1090 }, { "epoch": 0.7307434695244475, "grad_norm": 2.0659120011902545, "learning_rate": 4.896890162172731e-06, "loss": 1.6514, "step": 1091 }, { "epoch": 0.7314132618888145, "grad_norm": 2.1966927555380598, "learning_rate": 4.896474356503802e-06, "loss": 1.7797, "step": 1092 }, { "epoch": 0.7320830542531815, "grad_norm": 2.3094724813747933, "learning_rate": 4.896057731847521e-06, "loss": 1.7535, "step": 1093 }, { "epoch": 0.7327528466175486, "grad_norm": 2.2878815123686302, "learning_rate": 4.895640288346267e-06, "loss": 1.852, "step": 1094 }, { "epoch": 0.7334226389819156, "grad_norm": 2.647494140986839, "learning_rate": 4.8952220261427016e-06, "loss": 1.6549, "step": 1095 }, { "epoch": 0.7340924313462827, "grad_norm": 4.43842360738611, "learning_rate": 4.894802945379763e-06, "loss": 1.7553, "step": 1096 }, { "epoch": 0.7347622237106497, "grad_norm": 2.930904991004768, "learning_rate": 4.894383046200671e-06, "loss": 1.8377, "step": 1097 }, { "epoch": 0.7354320160750167, "grad_norm": 3.0308466439048076, "learning_rate": 4.893962328748927e-06, "loss": 1.5367, "step": 1098 }, { "epoch": 0.7361018084393838, "grad_norm": 3.6039312215176293, "learning_rate": 4.893540793168306e-06, "loss": 1.4941, "step": 1099 }, { "epoch": 0.7367716008037508, "grad_norm": 2.716710836659653, "learning_rate": 4.893118439602869e-06, "loss": 1.8076, "step": 1100 }, { "epoch": 0.7374413931681179, "grad_norm": 2.4193591285974207, "learning_rate": 4.892695268196954e-06, "loss": 1.5893, "step": 1101 }, { "epoch": 0.7381111855324849, "grad_norm": 4.110657286016066, "learning_rate": 4.892271279095176e-06, "loss": 1.6889, "step": 1102 }, { "epoch": 0.7387809778968519, "grad_norm": 2.7652222763375676, "learning_rate": 4.891846472442436e-06, "loss": 1.7121, "step": 1103 }, { "epoch": 0.739450770261219, "grad_norm": 3.067793274327967, "learning_rate": 4.891420848383906e-06, "loss": 1.7448, "step": 1104 }, { "epoch": 0.7401205626255861, "grad_norm": 2.3069805617625825, "learning_rate": 4.890994407065045e-06, "loss": 1.8696, "step": 1105 }, { "epoch": 0.7407903549899532, "grad_norm": 2.5149948833003473, "learning_rate": 4.890567148631587e-06, "loss": 1.5484, "step": 1106 }, { "epoch": 0.7414601473543202, "grad_norm": 2.3282422347310647, "learning_rate": 4.890139073229545e-06, "loss": 1.8104, "step": 1107 }, { "epoch": 0.7421299397186872, "grad_norm": 3.072309651491222, "learning_rate": 4.889710181005213e-06, "loss": 1.6801, "step": 1108 }, { "epoch": 0.7427997320830543, "grad_norm": 3.3350376225002782, "learning_rate": 4.889280472105164e-06, "loss": 1.8046, "step": 1109 }, { "epoch": 0.7434695244474213, "grad_norm": 2.517889204743094, "learning_rate": 4.8888499466762485e-06, "loss": 1.6247, "step": 1110 }, { "epoch": 0.7441393168117884, "grad_norm": 2.977878604665513, "learning_rate": 4.888418604865599e-06, "loss": 1.4777, "step": 1111 }, { "epoch": 0.7448091091761554, "grad_norm": 2.222067134317306, "learning_rate": 4.887986446820624e-06, "loss": 1.6506, "step": 1112 }, { "epoch": 0.7454789015405224, "grad_norm": 2.7900245491150586, "learning_rate": 4.8875534726890115e-06, "loss": 1.6701, "step": 1113 }, { "epoch": 0.7461486939048895, "grad_norm": 2.4078060975590048, "learning_rate": 4.887119682618731e-06, "loss": 1.6153, "step": 1114 }, { "epoch": 0.7468184862692565, "grad_norm": 3.275073530676623, "learning_rate": 4.8866850767580265e-06, "loss": 1.5926, "step": 1115 }, { "epoch": 0.7474882786336235, "grad_norm": 2.206920450254859, "learning_rate": 4.8862496552554255e-06, "loss": 1.7692, "step": 1116 }, { "epoch": 0.7481580709979906, "grad_norm": 2.115925700710982, "learning_rate": 4.885813418259729e-06, "loss": 1.849, "step": 1117 }, { "epoch": 0.7488278633623576, "grad_norm": 2.793742132169831, "learning_rate": 4.885376365920023e-06, "loss": 1.6506, "step": 1118 }, { "epoch": 0.7494976557267247, "grad_norm": 2.2307396679321743, "learning_rate": 4.884938498385665e-06, "loss": 1.8916, "step": 1119 }, { "epoch": 0.7501674480910918, "grad_norm": 3.3388438968815106, "learning_rate": 4.8844998158062986e-06, "loss": 1.7701, "step": 1120 }, { "epoch": 0.7508372404554589, "grad_norm": 2.195552455828595, "learning_rate": 4.884060318331839e-06, "loss": 1.5811, "step": 1121 }, { "epoch": 0.7515070328198259, "grad_norm": 4.012276424124285, "learning_rate": 4.883620006112485e-06, "loss": 1.6796, "step": 1122 }, { "epoch": 0.7521768251841929, "grad_norm": 2.842875284648362, "learning_rate": 4.8831788792987115e-06, "loss": 1.6365, "step": 1123 }, { "epoch": 0.75284661754856, "grad_norm": 5.545242300985668, "learning_rate": 4.8827369380412715e-06, "loss": 1.5135, "step": 1124 }, { "epoch": 0.753516409912927, "grad_norm": 3.240139611137648, "learning_rate": 4.882294182491198e-06, "loss": 1.7437, "step": 1125 }, { "epoch": 0.754186202277294, "grad_norm": 2.2248566481704937, "learning_rate": 4.8818506127998e-06, "loss": 1.7505, "step": 1126 }, { "epoch": 0.7548559946416611, "grad_norm": 2.089133581215171, "learning_rate": 4.881406229118667e-06, "loss": 1.7894, "step": 1127 }, { "epoch": 0.7555257870060281, "grad_norm": 2.765802896761999, "learning_rate": 4.880961031599665e-06, "loss": 1.6213, "step": 1128 }, { "epoch": 0.7561955793703952, "grad_norm": 3.9254586750328047, "learning_rate": 4.880515020394939e-06, "loss": 1.5351, "step": 1129 }, { "epoch": 0.7568653717347622, "grad_norm": 3.649323397012546, "learning_rate": 4.880068195656913e-06, "loss": 1.8637, "step": 1130 }, { "epoch": 0.7575351640991292, "grad_norm": 3.5949000079971194, "learning_rate": 4.879620557538286e-06, "loss": 1.8174, "step": 1131 }, { "epoch": 0.7582049564634963, "grad_norm": 2.1758632536477696, "learning_rate": 4.8791721061920385e-06, "loss": 1.5957, "step": 1132 }, { "epoch": 0.7588747488278633, "grad_norm": 3.104482768224697, "learning_rate": 4.8787228417714265e-06, "loss": 1.6068, "step": 1133 }, { "epoch": 0.7595445411922304, "grad_norm": 3.2802215842416573, "learning_rate": 4.878272764429985e-06, "loss": 1.5824, "step": 1134 }, { "epoch": 0.7602143335565975, "grad_norm": 2.492497692163386, "learning_rate": 4.877821874321526e-06, "loss": 1.8685, "step": 1135 }, { "epoch": 0.7608841259209645, "grad_norm": 3.5536658295690584, "learning_rate": 4.87737017160014e-06, "loss": 1.7572, "step": 1136 }, { "epoch": 0.7615539182853316, "grad_norm": 3.842658744535903, "learning_rate": 4.876917656420195e-06, "loss": 1.7338, "step": 1137 }, { "epoch": 0.7622237106496986, "grad_norm": 2.514055221515726, "learning_rate": 4.8764643289363365e-06, "loss": 1.5524, "step": 1138 }, { "epoch": 0.7628935030140657, "grad_norm": 2.4692602834788158, "learning_rate": 4.876010189303487e-06, "loss": 1.5575, "step": 1139 }, { "epoch": 0.7635632953784327, "grad_norm": 2.441943466239323, "learning_rate": 4.875555237676849e-06, "loss": 1.5661, "step": 1140 }, { "epoch": 0.7642330877427997, "grad_norm": 5.107982021520253, "learning_rate": 4.875099474211899e-06, "loss": 1.7533, "step": 1141 }, { "epoch": 0.7649028801071668, "grad_norm": 2.9437521268305042, "learning_rate": 4.874642899064395e-06, "loss": 1.6784, "step": 1142 }, { "epoch": 0.7655726724715338, "grad_norm": 2.541737708985127, "learning_rate": 4.874185512390367e-06, "loss": 1.8463, "step": 1143 }, { "epoch": 0.7662424648359009, "grad_norm": 2.9398310496704156, "learning_rate": 4.8737273143461275e-06, "loss": 1.6259, "step": 1144 }, { "epoch": 0.7669122572002679, "grad_norm": 4.198763580034348, "learning_rate": 4.8732683050882635e-06, "loss": 1.8523, "step": 1145 }, { "epoch": 0.7675820495646349, "grad_norm": 5.326104718673747, "learning_rate": 4.87280848477364e-06, "loss": 1.8367, "step": 1146 }, { "epoch": 0.768251841929002, "grad_norm": 2.389978219954923, "learning_rate": 4.872347853559399e-06, "loss": 1.7594, "step": 1147 }, { "epoch": 0.768921634293369, "grad_norm": 2.530052182627442, "learning_rate": 4.871886411602961e-06, "loss": 1.9483, "step": 1148 }, { "epoch": 0.769591426657736, "grad_norm": 2.142148315100061, "learning_rate": 4.8714241590620195e-06, "loss": 1.8082, "step": 1149 }, { "epoch": 0.7702612190221031, "grad_norm": 2.2251787197326434, "learning_rate": 4.870961096094551e-06, "loss": 1.6084, "step": 1150 }, { "epoch": 0.7709310113864702, "grad_norm": 2.567767785962183, "learning_rate": 4.870497222858803e-06, "loss": 1.7372, "step": 1151 }, { "epoch": 0.7716008037508373, "grad_norm": 6.523653621186929, "learning_rate": 4.870032539513305e-06, "loss": 1.7998, "step": 1152 }, { "epoch": 0.7722705961152043, "grad_norm": 4.411007062787531, "learning_rate": 4.869567046216859e-06, "loss": 1.2672, "step": 1153 }, { "epoch": 0.7729403884795714, "grad_norm": 3.4905317495846666, "learning_rate": 4.869100743128548e-06, "loss": 1.7845, "step": 1154 }, { "epoch": 0.7736101808439384, "grad_norm": 2.7385791852897907, "learning_rate": 4.868633630407727e-06, "loss": 1.8384, "step": 1155 }, { "epoch": 0.7742799732083054, "grad_norm": 2.6864528499847036, "learning_rate": 4.8681657082140325e-06, "loss": 1.8485, "step": 1156 }, { "epoch": 0.7749497655726725, "grad_norm": 4.58170107036923, "learning_rate": 4.867696976707373e-06, "loss": 1.6156, "step": 1157 }, { "epoch": 0.7756195579370395, "grad_norm": 2.316586910268411, "learning_rate": 4.867227436047939e-06, "loss": 1.763, "step": 1158 }, { "epoch": 0.7762893503014066, "grad_norm": 2.125613072998781, "learning_rate": 4.866757086396193e-06, "loss": 1.8737, "step": 1159 }, { "epoch": 0.7769591426657736, "grad_norm": 2.1895715988415314, "learning_rate": 4.866285927912875e-06, "loss": 1.8214, "step": 1160 }, { "epoch": 0.7776289350301406, "grad_norm": 3.754339108388947, "learning_rate": 4.865813960759002e-06, "loss": 1.897, "step": 1161 }, { "epoch": 0.7782987273945077, "grad_norm": 2.234039467856894, "learning_rate": 4.8653411850958685e-06, "loss": 1.5274, "step": 1162 }, { "epoch": 0.7789685197588747, "grad_norm": 6.078190998384283, "learning_rate": 4.8648676010850424e-06, "loss": 1.4745, "step": 1163 }, { "epoch": 0.7796383121232418, "grad_norm": 3.0904620688931206, "learning_rate": 4.864393208888371e-06, "loss": 1.5436, "step": 1164 }, { "epoch": 0.7803081044876088, "grad_norm": 2.760444510476112, "learning_rate": 4.863918008667975e-06, "loss": 1.5259, "step": 1165 }, { "epoch": 0.7809778968519759, "grad_norm": 2.5661214599480817, "learning_rate": 4.863442000586253e-06, "loss": 1.6754, "step": 1166 }, { "epoch": 0.781647689216343, "grad_norm": 2.3973732359819686, "learning_rate": 4.8629651848058805e-06, "loss": 1.6797, "step": 1167 }, { "epoch": 0.78231748158071, "grad_norm": 2.2195463269898124, "learning_rate": 4.862487561489805e-06, "loss": 1.765, "step": 1168 }, { "epoch": 0.7829872739450771, "grad_norm": 2.0928176846816235, "learning_rate": 4.862009130801255e-06, "loss": 1.7553, "step": 1169 }, { "epoch": 0.7836570663094441, "grad_norm": 2.12478530144836, "learning_rate": 4.861529892903731e-06, "loss": 1.8533, "step": 1170 }, { "epoch": 0.7843268586738111, "grad_norm": 2.2040019801703727, "learning_rate": 4.861049847961014e-06, "loss": 1.7366, "step": 1171 }, { "epoch": 0.7849966510381782, "grad_norm": 3.6742505362085076, "learning_rate": 4.8605689961371536e-06, "loss": 1.6808, "step": 1172 }, { "epoch": 0.7856664434025452, "grad_norm": 2.218091233374028, "learning_rate": 4.860087337596482e-06, "loss": 1.7187, "step": 1173 }, { "epoch": 0.7863362357669123, "grad_norm": 2.34064451146891, "learning_rate": 4.859604872503604e-06, "loss": 1.8842, "step": 1174 }, { "epoch": 0.7870060281312793, "grad_norm": 3.107160391923673, "learning_rate": 4.859121601023401e-06, "loss": 1.8086, "step": 1175 }, { "epoch": 0.7876758204956463, "grad_norm": 2.1794307334328473, "learning_rate": 4.858637523321028e-06, "loss": 1.643, "step": 1176 }, { "epoch": 0.7883456128600134, "grad_norm": 2.1283239244792314, "learning_rate": 4.8581526395619184e-06, "loss": 1.6467, "step": 1177 }, { "epoch": 0.7890154052243804, "grad_norm": 3.1344465280761997, "learning_rate": 4.857666949911779e-06, "loss": 1.8454, "step": 1178 }, { "epoch": 0.7896851975887474, "grad_norm": 7.622625714347746, "learning_rate": 4.857180454536593e-06, "loss": 1.7255, "step": 1179 }, { "epoch": 0.7903549899531145, "grad_norm": 2.6452074603976103, "learning_rate": 4.856693153602618e-06, "loss": 1.6269, "step": 1180 }, { "epoch": 0.7910247823174816, "grad_norm": 2.7016749636413393, "learning_rate": 4.856205047276388e-06, "loss": 1.7076, "step": 1181 }, { "epoch": 0.7916945746818487, "grad_norm": 2.285343790337882, "learning_rate": 4.855716135724712e-06, "loss": 1.7405, "step": 1182 }, { "epoch": 0.7923643670462157, "grad_norm": 5.72117605034383, "learning_rate": 4.855226419114673e-06, "loss": 1.3304, "step": 1183 }, { "epoch": 0.7930341594105828, "grad_norm": 8.552804638493493, "learning_rate": 4.854735897613631e-06, "loss": 1.6013, "step": 1184 }, { "epoch": 0.7937039517749498, "grad_norm": 2.165414766351289, "learning_rate": 4.85424457138922e-06, "loss": 1.6853, "step": 1185 }, { "epoch": 0.7943737441393168, "grad_norm": 2.01682652371573, "learning_rate": 4.853752440609348e-06, "loss": 1.6803, "step": 1186 }, { "epoch": 0.7950435365036839, "grad_norm": 3.9582959130491737, "learning_rate": 4.853259505442202e-06, "loss": 1.821, "step": 1187 }, { "epoch": 0.7957133288680509, "grad_norm": 2.620473932162896, "learning_rate": 4.852765766056238e-06, "loss": 1.7646, "step": 1188 }, { "epoch": 0.796383121232418, "grad_norm": 2.1014311962630936, "learning_rate": 4.85227122262019e-06, "loss": 1.5884, "step": 1189 }, { "epoch": 0.797052913596785, "grad_norm": 2.386958014593894, "learning_rate": 4.851775875303069e-06, "loss": 1.5605, "step": 1190 }, { "epoch": 0.797722705961152, "grad_norm": 3.20154552521541, "learning_rate": 4.851279724274155e-06, "loss": 1.6721, "step": 1191 }, { "epoch": 0.7983924983255191, "grad_norm": 3.601320621855959, "learning_rate": 4.850782769703009e-06, "loss": 1.6764, "step": 1192 }, { "epoch": 0.7990622906898861, "grad_norm": 4.37769950925315, "learning_rate": 4.850285011759462e-06, "loss": 1.5573, "step": 1193 }, { "epoch": 0.7997320830542531, "grad_norm": 2.3149974951453345, "learning_rate": 4.849786450613622e-06, "loss": 1.8865, "step": 1194 }, { "epoch": 0.8004018754186202, "grad_norm": 5.225102876303946, "learning_rate": 4.849287086435871e-06, "loss": 1.7707, "step": 1195 }, { "epoch": 0.8010716677829873, "grad_norm": 2.111365076268949, "learning_rate": 4.848786919396864e-06, "loss": 1.6565, "step": 1196 }, { "epoch": 0.8017414601473544, "grad_norm": 2.5972435765883217, "learning_rate": 4.848285949667532e-06, "loss": 1.5702, "step": 1197 }, { "epoch": 0.8024112525117214, "grad_norm": 2.5347008278595, "learning_rate": 4.847784177419079e-06, "loss": 1.6771, "step": 1198 }, { "epoch": 0.8030810448760884, "grad_norm": 3.2950554556606018, "learning_rate": 4.847281602822985e-06, "loss": 1.5734, "step": 1199 }, { "epoch": 0.8037508372404555, "grad_norm": 7.825483330773524, "learning_rate": 4.846778226051002e-06, "loss": 1.4391, "step": 1200 }, { "epoch": 0.8044206296048225, "grad_norm": 2.252434143192482, "learning_rate": 4.846274047275159e-06, "loss": 1.4057, "step": 1201 }, { "epoch": 0.8050904219691896, "grad_norm": 4.227011790532466, "learning_rate": 4.845769066667757e-06, "loss": 1.6178, "step": 1202 }, { "epoch": 0.8057602143335566, "grad_norm": 4.443950712155468, "learning_rate": 4.845263284401371e-06, "loss": 1.7194, "step": 1203 }, { "epoch": 0.8064300066979236, "grad_norm": 2.432547037965096, "learning_rate": 4.844756700648852e-06, "loss": 1.6696, "step": 1204 }, { "epoch": 0.8070997990622907, "grad_norm": 4.659510910458834, "learning_rate": 4.844249315583321e-06, "loss": 1.7285, "step": 1205 }, { "epoch": 0.8077695914266577, "grad_norm": 2.600586693976699, "learning_rate": 4.843741129378175e-06, "loss": 1.6138, "step": 1206 }, { "epoch": 0.8084393837910248, "grad_norm": 2.4470121490612744, "learning_rate": 4.843232142207088e-06, "loss": 1.838, "step": 1207 }, { "epoch": 0.8091091761553918, "grad_norm": 2.2840724156610994, "learning_rate": 4.842722354244001e-06, "loss": 1.6826, "step": 1208 }, { "epoch": 0.8097789685197588, "grad_norm": 2.1891621930414127, "learning_rate": 4.842211765663136e-06, "loss": 1.5532, "step": 1209 }, { "epoch": 0.8104487608841259, "grad_norm": 2.4627704747242576, "learning_rate": 4.841700376638983e-06, "loss": 1.5835, "step": 1210 }, { "epoch": 0.8111185532484929, "grad_norm": 3.581612655850042, "learning_rate": 4.8411881873463085e-06, "loss": 1.781, "step": 1211 }, { "epoch": 0.8117883456128601, "grad_norm": 3.3574341134159345, "learning_rate": 4.840675197960151e-06, "loss": 1.8507, "step": 1212 }, { "epoch": 0.8124581379772271, "grad_norm": 4.143232943871543, "learning_rate": 4.840161408655822e-06, "loss": 1.7192, "step": 1213 }, { "epoch": 0.8131279303415941, "grad_norm": 3.9511750594793766, "learning_rate": 4.839646819608908e-06, "loss": 1.5458, "step": 1214 }, { "epoch": 0.8137977227059612, "grad_norm": 2.322395284280297, "learning_rate": 4.8391314309952695e-06, "loss": 1.8197, "step": 1215 }, { "epoch": 0.8144675150703282, "grad_norm": 2.494421265501521, "learning_rate": 4.838615242991037e-06, "loss": 1.6384, "step": 1216 }, { "epoch": 0.8151373074346953, "grad_norm": 2.175915324875781, "learning_rate": 4.838098255772617e-06, "loss": 1.6858, "step": 1217 }, { "epoch": 0.8158070997990623, "grad_norm": 5.518379949315218, "learning_rate": 4.837580469516688e-06, "loss": 1.7574, "step": 1218 }, { "epoch": 0.8164768921634293, "grad_norm": 2.6515456516363627, "learning_rate": 4.837061884400202e-06, "loss": 1.6605, "step": 1219 }, { "epoch": 0.8171466845277964, "grad_norm": 5.423838772077017, "learning_rate": 4.836542500600383e-06, "loss": 1.5525, "step": 1220 }, { "epoch": 0.8178164768921634, "grad_norm": 6.354480440704069, "learning_rate": 4.83602231829473e-06, "loss": 1.7938, "step": 1221 }, { "epoch": 0.8184862692565305, "grad_norm": 2.2263387186428267, "learning_rate": 4.8355013376610124e-06, "loss": 1.6579, "step": 1222 }, { "epoch": 0.8191560616208975, "grad_norm": 3.616932296156974, "learning_rate": 4.834979558877274e-06, "loss": 1.623, "step": 1223 }, { "epoch": 0.8198258539852645, "grad_norm": 2.4897066530165977, "learning_rate": 4.834456982121832e-06, "loss": 1.8656, "step": 1224 }, { "epoch": 0.8204956463496316, "grad_norm": 2.15018609228008, "learning_rate": 4.8339336075732745e-06, "loss": 1.7469, "step": 1225 }, { "epoch": 0.8211654387139986, "grad_norm": 5.688676966099046, "learning_rate": 4.833409435410463e-06, "loss": 1.7943, "step": 1226 }, { "epoch": 0.8218352310783658, "grad_norm": 2.5229462977616492, "learning_rate": 4.8328844658125305e-06, "loss": 1.768, "step": 1227 }, { "epoch": 0.8225050234427328, "grad_norm": 3.567856130399705, "learning_rate": 4.832358698958887e-06, "loss": 1.5309, "step": 1228 }, { "epoch": 0.8231748158070998, "grad_norm": 2.4374889960066226, "learning_rate": 4.831832135029208e-06, "loss": 1.8005, "step": 1229 }, { "epoch": 0.8238446081714669, "grad_norm": 3.0201438888746277, "learning_rate": 4.831304774203447e-06, "loss": 1.8235, "step": 1230 }, { "epoch": 0.8245144005358339, "grad_norm": 2.2251756509173486, "learning_rate": 4.830776616661827e-06, "loss": 1.7069, "step": 1231 }, { "epoch": 0.825184192900201, "grad_norm": 2.249068103076129, "learning_rate": 4.830247662584845e-06, "loss": 1.7793, "step": 1232 }, { "epoch": 0.825853985264568, "grad_norm": 2.8513375215659904, "learning_rate": 4.829717912153268e-06, "loss": 1.6882, "step": 1233 }, { "epoch": 0.826523777628935, "grad_norm": 2.253290610216783, "learning_rate": 4.829187365548138e-06, "loss": 1.7198, "step": 1234 }, { "epoch": 0.8271935699933021, "grad_norm": 2.5514091094324027, "learning_rate": 4.828656022950767e-06, "loss": 1.5131, "step": 1235 }, { "epoch": 0.8278633623576691, "grad_norm": 2.3468149559822034, "learning_rate": 4.828123884542739e-06, "loss": 1.7255, "step": 1236 }, { "epoch": 0.8285331547220361, "grad_norm": 3.912263725057324, "learning_rate": 4.827590950505911e-06, "loss": 1.6795, "step": 1237 }, { "epoch": 0.8292029470864032, "grad_norm": 2.193900559498398, "learning_rate": 4.827057221022413e-06, "loss": 1.6858, "step": 1238 }, { "epoch": 0.8298727394507702, "grad_norm": 2.552293145704249, "learning_rate": 4.826522696274642e-06, "loss": 1.6984, "step": 1239 }, { "epoch": 0.8305425318151373, "grad_norm": 4.180207940538332, "learning_rate": 4.825987376445273e-06, "loss": 1.5398, "step": 1240 }, { "epoch": 0.8312123241795043, "grad_norm": 2.58031146942115, "learning_rate": 4.82545126171725e-06, "loss": 1.6417, "step": 1241 }, { "epoch": 0.8318821165438715, "grad_norm": 2.1920157450508597, "learning_rate": 4.824914352273786e-06, "loss": 1.5449, "step": 1242 }, { "epoch": 0.8325519089082385, "grad_norm": 3.84893930485421, "learning_rate": 4.82437664829837e-06, "loss": 1.6588, "step": 1243 }, { "epoch": 0.8332217012726055, "grad_norm": 2.149908802999939, "learning_rate": 4.823838149974759e-06, "loss": 1.75, "step": 1244 }, { "epoch": 0.8338914936369726, "grad_norm": 2.956846048067086, "learning_rate": 4.823298857486984e-06, "loss": 1.5346, "step": 1245 }, { "epoch": 0.8345612860013396, "grad_norm": 3.892370518708501, "learning_rate": 4.822758771019348e-06, "loss": 1.5701, "step": 1246 }, { "epoch": 0.8352310783657066, "grad_norm": 2.87772119338136, "learning_rate": 4.822217890756421e-06, "loss": 1.7883, "step": 1247 }, { "epoch": 0.8359008707300737, "grad_norm": 2.12704574882012, "learning_rate": 4.821676216883049e-06, "loss": 1.6245, "step": 1248 }, { "epoch": 0.8365706630944407, "grad_norm": 2.258683925718426, "learning_rate": 4.821133749584346e-06, "loss": 1.6803, "step": 1249 }, { "epoch": 0.8372404554588078, "grad_norm": 4.633866146034471, "learning_rate": 4.820590489045699e-06, "loss": 1.6706, "step": 1250 }, { "epoch": 0.8379102478231748, "grad_norm": 2.796753057951858, "learning_rate": 4.820046435452767e-06, "loss": 1.651, "step": 1251 }, { "epoch": 0.8385800401875418, "grad_norm": 2.637617705360298, "learning_rate": 4.819501588991475e-06, "loss": 1.8155, "step": 1252 }, { "epoch": 0.8392498325519089, "grad_norm": 3.9272273695588424, "learning_rate": 4.818955949848027e-06, "loss": 1.7201, "step": 1253 }, { "epoch": 0.8399196249162759, "grad_norm": 2.9077382508033263, "learning_rate": 4.81840951820889e-06, "loss": 1.4502, "step": 1254 }, { "epoch": 0.840589417280643, "grad_norm": 2.621067798620622, "learning_rate": 4.817862294260806e-06, "loss": 1.6703, "step": 1255 }, { "epoch": 0.84125920964501, "grad_norm": 2.331164054504543, "learning_rate": 4.817314278190788e-06, "loss": 1.6763, "step": 1256 }, { "epoch": 0.8419290020093771, "grad_norm": 2.174458261624428, "learning_rate": 4.816765470186117e-06, "loss": 1.8009, "step": 1257 }, { "epoch": 0.8425987943737442, "grad_norm": 2.1815805111734377, "learning_rate": 4.816215870434349e-06, "loss": 1.4291, "step": 1258 }, { "epoch": 0.8432685867381112, "grad_norm": 2.113277445914001, "learning_rate": 4.815665479123306e-06, "loss": 1.8376, "step": 1259 }, { "epoch": 0.8439383791024783, "grad_norm": 3.896943091206253, "learning_rate": 4.815114296441082e-06, "loss": 1.5013, "step": 1260 }, { "epoch": 0.8446081714668453, "grad_norm": 2.0702483792777224, "learning_rate": 4.814562322576043e-06, "loss": 1.6789, "step": 1261 }, { "epoch": 0.8452779638312123, "grad_norm": 2.263936674620398, "learning_rate": 4.814009557716825e-06, "loss": 1.6884, "step": 1262 }, { "epoch": 0.8459477561955794, "grad_norm": 2.2593770661746966, "learning_rate": 4.813456002052331e-06, "loss": 1.8365, "step": 1263 }, { "epoch": 0.8466175485599464, "grad_norm": 2.5832353097570584, "learning_rate": 4.81290165577174e-06, "loss": 1.6678, "step": 1264 }, { "epoch": 0.8472873409243135, "grad_norm": 2.435374885771977, "learning_rate": 4.812346519064496e-06, "loss": 1.6581, "step": 1265 }, { "epoch": 0.8479571332886805, "grad_norm": 2.568216278987749, "learning_rate": 4.811790592120316e-06, "loss": 1.6529, "step": 1266 }, { "epoch": 0.8486269256530475, "grad_norm": 2.2018738811567586, "learning_rate": 4.811233875129185e-06, "loss": 1.7218, "step": 1267 }, { "epoch": 0.8492967180174146, "grad_norm": 2.2568104505820687, "learning_rate": 4.810676368281361e-06, "loss": 1.6202, "step": 1268 }, { "epoch": 0.8499665103817816, "grad_norm": 3.722099021568899, "learning_rate": 4.8101180717673686e-06, "loss": 1.68, "step": 1269 }, { "epoch": 0.8506363027461487, "grad_norm": 3.0715171963975805, "learning_rate": 4.809558985778004e-06, "loss": 1.6874, "step": 1270 }, { "epoch": 0.8513060951105157, "grad_norm": 2.4884371818682953, "learning_rate": 4.808999110504335e-06, "loss": 1.7379, "step": 1271 }, { "epoch": 0.8519758874748827, "grad_norm": 9.310893768806528, "learning_rate": 4.808438446137696e-06, "loss": 1.6855, "step": 1272 }, { "epoch": 0.8526456798392499, "grad_norm": 4.567066689779984, "learning_rate": 4.80787699286969e-06, "loss": 1.8022, "step": 1273 }, { "epoch": 0.8533154722036169, "grad_norm": 3.286249286911338, "learning_rate": 4.807314750892195e-06, "loss": 1.7806, "step": 1274 }, { "epoch": 0.853985264567984, "grad_norm": 4.968724407082144, "learning_rate": 4.806751720397354e-06, "loss": 1.5724, "step": 1275 }, { "epoch": 0.854655056932351, "grad_norm": 3.6928212252293, "learning_rate": 4.806187901577581e-06, "loss": 1.5573, "step": 1276 }, { "epoch": 0.855324849296718, "grad_norm": 4.119450870352165, "learning_rate": 4.8056232946255595e-06, "loss": 1.5087, "step": 1277 }, { "epoch": 0.8559946416610851, "grad_norm": 2.253454682225228, "learning_rate": 4.805057899734241e-06, "loss": 1.8089, "step": 1278 }, { "epoch": 0.8566644340254521, "grad_norm": 2.202306129444853, "learning_rate": 4.8044917170968496e-06, "loss": 1.7437, "step": 1279 }, { "epoch": 0.8573342263898192, "grad_norm": 2.314589060601397, "learning_rate": 4.803924746906874e-06, "loss": 1.668, "step": 1280 }, { "epoch": 0.8580040187541862, "grad_norm": 2.1843795958446615, "learning_rate": 4.803356989358076e-06, "loss": 1.8496, "step": 1281 }, { "epoch": 0.8586738111185532, "grad_norm": 2.3568108762077014, "learning_rate": 4.802788444644484e-06, "loss": 1.8163, "step": 1282 }, { "epoch": 0.8593436034829203, "grad_norm": 3.674019041475965, "learning_rate": 4.802219112960397e-06, "loss": 1.5976, "step": 1283 }, { "epoch": 0.8600133958472873, "grad_norm": 2.4669654583047276, "learning_rate": 4.801648994500382e-06, "loss": 1.5622, "step": 1284 }, { "epoch": 0.8606831882116543, "grad_norm": 2.3387702098298297, "learning_rate": 4.801078089459275e-06, "loss": 1.6788, "step": 1285 }, { "epoch": 0.8613529805760214, "grad_norm": 3.6815724605619584, "learning_rate": 4.800506398032179e-06, "loss": 1.624, "step": 1286 }, { "epoch": 0.8620227729403884, "grad_norm": 2.2458530700990837, "learning_rate": 4.7999339204144725e-06, "loss": 1.7261, "step": 1287 }, { "epoch": 0.8626925653047556, "grad_norm": 4.381510325718988, "learning_rate": 4.799360656801794e-06, "loss": 1.6855, "step": 1288 }, { "epoch": 0.8633623576691226, "grad_norm": 4.459243085556426, "learning_rate": 4.798786607390056e-06, "loss": 1.8165, "step": 1289 }, { "epoch": 0.8640321500334897, "grad_norm": 5.787695119201999, "learning_rate": 4.798211772375437e-06, "loss": 1.7344, "step": 1290 }, { "epoch": 0.8647019423978567, "grad_norm": 3.52992823632776, "learning_rate": 4.797636151954386e-06, "loss": 1.636, "step": 1291 }, { "epoch": 0.8653717347622237, "grad_norm": 2.349535082622118, "learning_rate": 4.797059746323619e-06, "loss": 1.8238, "step": 1292 }, { "epoch": 0.8660415271265908, "grad_norm": 2.393827527090311, "learning_rate": 4.796482555680121e-06, "loss": 1.4413, "step": 1293 }, { "epoch": 0.8667113194909578, "grad_norm": 7.5002901792324, "learning_rate": 4.795904580221144e-06, "loss": 1.599, "step": 1294 }, { "epoch": 0.8673811118553248, "grad_norm": 3.991034425878605, "learning_rate": 4.795325820144211e-06, "loss": 1.4894, "step": 1295 }, { "epoch": 0.8680509042196919, "grad_norm": 5.755389745572441, "learning_rate": 4.79474627564711e-06, "loss": 1.7081, "step": 1296 }, { "epoch": 0.8687206965840589, "grad_norm": 2.077866241848505, "learning_rate": 4.794165946927898e-06, "loss": 1.5718, "step": 1297 }, { "epoch": 0.869390488948426, "grad_norm": 3.1495418790854823, "learning_rate": 4.793584834184902e-06, "loss": 1.627, "step": 1298 }, { "epoch": 0.870060281312793, "grad_norm": 2.7614490392529305, "learning_rate": 4.793002937616715e-06, "loss": 1.5924, "step": 1299 }, { "epoch": 0.87073007367716, "grad_norm": 3.154027259753376, "learning_rate": 4.792420257422199e-06, "loss": 1.3415, "step": 1300 }, { "epoch": 0.8713998660415271, "grad_norm": 2.5585142544343715, "learning_rate": 4.79183679380048e-06, "loss": 1.6787, "step": 1301 }, { "epoch": 0.8720696584058941, "grad_norm": 3.94896999173285, "learning_rate": 4.791252546950959e-06, "loss": 1.7519, "step": 1302 }, { "epoch": 0.8727394507702613, "grad_norm": 2.224652572648891, "learning_rate": 4.790667517073297e-06, "loss": 1.7537, "step": 1303 }, { "epoch": 0.8734092431346283, "grad_norm": 3.007989714325827, "learning_rate": 4.7900817043674274e-06, "loss": 1.5614, "step": 1304 }, { "epoch": 0.8740790354989953, "grad_norm": 2.553721939769534, "learning_rate": 4.78949510903355e-06, "loss": 1.6467, "step": 1305 }, { "epoch": 0.8747488278633624, "grad_norm": 3.443496754613428, "learning_rate": 4.788907731272132e-06, "loss": 1.6137, "step": 1306 }, { "epoch": 0.8754186202277294, "grad_norm": 2.7521270338150083, "learning_rate": 4.788319571283907e-06, "loss": 1.3832, "step": 1307 }, { "epoch": 0.8760884125920965, "grad_norm": 2.3384351894395374, "learning_rate": 4.787730629269877e-06, "loss": 1.7708, "step": 1308 }, { "epoch": 0.8767582049564635, "grad_norm": 2.60924381876655, "learning_rate": 4.7871409054313125e-06, "loss": 1.6922, "step": 1309 }, { "epoch": 0.8774279973208305, "grad_norm": 2.5844781467284768, "learning_rate": 4.786550399969748e-06, "loss": 1.5847, "step": 1310 }, { "epoch": 0.8780977896851976, "grad_norm": 2.3028852045502903, "learning_rate": 4.7859591130869876e-06, "loss": 1.5768, "step": 1311 }, { "epoch": 0.8787675820495646, "grad_norm": 2.097221500047137, "learning_rate": 4.785367044985101e-06, "loss": 1.6672, "step": 1312 }, { "epoch": 0.8794373744139317, "grad_norm": 2.071197944099126, "learning_rate": 4.784774195866427e-06, "loss": 1.8165, "step": 1313 }, { "epoch": 0.8801071667782987, "grad_norm": 2.529616628741377, "learning_rate": 4.784180565933567e-06, "loss": 1.7586, "step": 1314 }, { "epoch": 0.8807769591426657, "grad_norm": 3.504578526489725, "learning_rate": 4.7835861553893955e-06, "loss": 1.7369, "step": 1315 }, { "epoch": 0.8814467515070328, "grad_norm": 2.544154753331578, "learning_rate": 4.782990964437048e-06, "loss": 1.656, "step": 1316 }, { "epoch": 0.8821165438713998, "grad_norm": 7.467725672441741, "learning_rate": 4.782394993279931e-06, "loss": 1.6514, "step": 1317 }, { "epoch": 0.882786336235767, "grad_norm": 2.739716841924104, "learning_rate": 4.7817982421217134e-06, "loss": 1.6772, "step": 1318 }, { "epoch": 0.883456128600134, "grad_norm": 2.3394307205306237, "learning_rate": 4.781200711166335e-06, "loss": 1.6398, "step": 1319 }, { "epoch": 0.884125920964501, "grad_norm": 2.306554395134671, "learning_rate": 4.780602400617999e-06, "loss": 1.7233, "step": 1320 }, { "epoch": 0.8847957133288681, "grad_norm": 2.3060665298056966, "learning_rate": 4.780003310681177e-06, "loss": 1.6264, "step": 1321 }, { "epoch": 0.8854655056932351, "grad_norm": 2.652934997810156, "learning_rate": 4.779403441560604e-06, "loss": 1.7401, "step": 1322 }, { "epoch": 0.8861352980576022, "grad_norm": 2.230433297825902, "learning_rate": 4.778802793461287e-06, "loss": 1.5142, "step": 1323 }, { "epoch": 0.8868050904219692, "grad_norm": 2.2762551373236652, "learning_rate": 4.778201366588491e-06, "loss": 1.6628, "step": 1324 }, { "epoch": 0.8874748827863362, "grad_norm": 2.1552430306753143, "learning_rate": 4.777599161147755e-06, "loss": 1.7375, "step": 1325 }, { "epoch": 0.8881446751507033, "grad_norm": 2.2607727494640213, "learning_rate": 4.77699617734488e-06, "loss": 1.6244, "step": 1326 }, { "epoch": 0.8888144675150703, "grad_norm": 3.837370403839305, "learning_rate": 4.7763924153859334e-06, "loss": 1.5682, "step": 1327 }, { "epoch": 0.8894842598794374, "grad_norm": 2.2274663102358243, "learning_rate": 4.775787875477248e-06, "loss": 1.5196, "step": 1328 }, { "epoch": 0.8901540522438044, "grad_norm": 2.616547189490284, "learning_rate": 4.775182557825426e-06, "loss": 1.4947, "step": 1329 }, { "epoch": 0.8908238446081714, "grad_norm": 2.409325713051335, "learning_rate": 4.77457646263733e-06, "loss": 1.7016, "step": 1330 }, { "epoch": 0.8914936369725385, "grad_norm": 2.175576794047721, "learning_rate": 4.773969590120092e-06, "loss": 1.5389, "step": 1331 }, { "epoch": 0.8921634293369055, "grad_norm": 2.2112331953983664, "learning_rate": 4.77336194048111e-06, "loss": 1.4761, "step": 1332 }, { "epoch": 0.8928332217012726, "grad_norm": 3.411255592667592, "learning_rate": 4.7727535139280446e-06, "loss": 1.6419, "step": 1333 }, { "epoch": 0.8935030140656397, "grad_norm": 2.3239028224291673, "learning_rate": 4.772144310668824e-06, "loss": 1.8766, "step": 1334 }, { "epoch": 0.8941728064300067, "grad_norm": 2.7452063332177032, "learning_rate": 4.771534330911641e-06, "loss": 1.6923, "step": 1335 }, { "epoch": 0.8948425987943738, "grad_norm": 2.095721841835635, "learning_rate": 4.770923574864957e-06, "loss": 1.6901, "step": 1336 }, { "epoch": 0.8955123911587408, "grad_norm": 2.206097337183002, "learning_rate": 4.770312042737491e-06, "loss": 1.7541, "step": 1337 }, { "epoch": 0.8961821835231079, "grad_norm": 4.4814953416928605, "learning_rate": 4.769699734738237e-06, "loss": 1.662, "step": 1338 }, { "epoch": 0.8968519758874749, "grad_norm": 2.336541040158993, "learning_rate": 4.769086651076447e-06, "loss": 1.8403, "step": 1339 }, { "epoch": 0.8975217682518419, "grad_norm": 3.27925035604721, "learning_rate": 4.768472791961639e-06, "loss": 1.767, "step": 1340 }, { "epoch": 0.898191560616209, "grad_norm": 2.831222202172219, "learning_rate": 4.7678581576036e-06, "loss": 1.3575, "step": 1341 }, { "epoch": 0.898861352980576, "grad_norm": 3.539811773314982, "learning_rate": 4.767242748212379e-06, "loss": 1.5624, "step": 1342 }, { "epoch": 0.899531145344943, "grad_norm": 2.847882166725228, "learning_rate": 4.766626563998288e-06, "loss": 1.7804, "step": 1343 }, { "epoch": 0.9002009377093101, "grad_norm": 2.1612802442240633, "learning_rate": 4.766009605171908e-06, "loss": 1.4386, "step": 1344 }, { "epoch": 0.9008707300736771, "grad_norm": 3.3939504028582936, "learning_rate": 4.765391871944081e-06, "loss": 1.7392, "step": 1345 }, { "epoch": 0.9015405224380442, "grad_norm": 2.5669610307188013, "learning_rate": 4.764773364525916e-06, "loss": 1.4744, "step": 1346 }, { "epoch": 0.9022103148024112, "grad_norm": 4.197578976800316, "learning_rate": 4.7641540831287884e-06, "loss": 1.5284, "step": 1347 }, { "epoch": 0.9028801071667782, "grad_norm": 2.941375822168656, "learning_rate": 4.763534027964332e-06, "loss": 1.4904, "step": 1348 }, { "epoch": 0.9035498995311454, "grad_norm": 2.778021975202989, "learning_rate": 4.76291319924445e-06, "loss": 1.6254, "step": 1349 }, { "epoch": 0.9042196918955124, "grad_norm": 2.2509555267859054, "learning_rate": 4.762291597181309e-06, "loss": 1.5045, "step": 1350 }, { "epoch": 0.9048894842598795, "grad_norm": 4.774133634592602, "learning_rate": 4.7616692219873394e-06, "loss": 1.5301, "step": 1351 }, { "epoch": 0.9055592766242465, "grad_norm": 4.103604714863219, "learning_rate": 4.761046073875235e-06, "loss": 1.732, "step": 1352 }, { "epoch": 0.9062290689886136, "grad_norm": 2.3859661607919005, "learning_rate": 4.760422153057956e-06, "loss": 1.6712, "step": 1353 }, { "epoch": 0.9068988613529806, "grad_norm": 2.1344449461378048, "learning_rate": 4.759797459748724e-06, "loss": 1.7901, "step": 1354 }, { "epoch": 0.9075686537173476, "grad_norm": 2.4027884002059863, "learning_rate": 4.759171994161027e-06, "loss": 1.6697, "step": 1355 }, { "epoch": 0.9082384460817147, "grad_norm": 3.195302940044241, "learning_rate": 4.758545756508616e-06, "loss": 1.7749, "step": 1356 }, { "epoch": 0.9089082384460817, "grad_norm": 3.376276807581927, "learning_rate": 4.757918747005504e-06, "loss": 1.688, "step": 1357 }, { "epoch": 0.9095780308104487, "grad_norm": 2.8076352567760217, "learning_rate": 4.757290965865971e-06, "loss": 1.7569, "step": 1358 }, { "epoch": 0.9102478231748158, "grad_norm": 4.812153048116083, "learning_rate": 4.756662413304559e-06, "loss": 1.5374, "step": 1359 }, { "epoch": 0.9109176155391828, "grad_norm": 3.508978108944886, "learning_rate": 4.756033089536073e-06, "loss": 1.6296, "step": 1360 }, { "epoch": 0.9115874079035499, "grad_norm": 2.656237709153644, "learning_rate": 4.755402994775583e-06, "loss": 1.6658, "step": 1361 }, { "epoch": 0.9122572002679169, "grad_norm": 2.270845669694639, "learning_rate": 4.754772129238423e-06, "loss": 1.7293, "step": 1362 }, { "epoch": 0.9129269926322839, "grad_norm": 7.733649931883227, "learning_rate": 4.754140493140187e-06, "loss": 1.6367, "step": 1363 }, { "epoch": 0.9135967849966511, "grad_norm": 3.7064142159263653, "learning_rate": 4.753508086696735e-06, "loss": 1.6278, "step": 1364 }, { "epoch": 0.9142665773610181, "grad_norm": 2.3186218861638896, "learning_rate": 4.752874910124191e-06, "loss": 1.7175, "step": 1365 }, { "epoch": 0.9149363697253852, "grad_norm": 2.2834657625023715, "learning_rate": 4.75224096363894e-06, "loss": 1.6459, "step": 1366 }, { "epoch": 0.9156061620897522, "grad_norm": 3.55715943201332, "learning_rate": 4.751606247457632e-06, "loss": 1.6842, "step": 1367 }, { "epoch": 0.9162759544541192, "grad_norm": 2.302965008303449, "learning_rate": 4.750970761797179e-06, "loss": 1.7211, "step": 1368 }, { "epoch": 0.9169457468184863, "grad_norm": 4.752085153271039, "learning_rate": 4.750334506874756e-06, "loss": 1.4983, "step": 1369 }, { "epoch": 0.9176155391828533, "grad_norm": 2.6078816992421037, "learning_rate": 4.749697482907801e-06, "loss": 1.6459, "step": 1370 }, { "epoch": 0.9182853315472204, "grad_norm": 1.9538229858664862, "learning_rate": 4.749059690114016e-06, "loss": 1.4483, "step": 1371 }, { "epoch": 0.9189551239115874, "grad_norm": 2.403254693862871, "learning_rate": 4.7484211287113625e-06, "loss": 1.8322, "step": 1372 }, { "epoch": 0.9196249162759544, "grad_norm": 2.1149407544575807, "learning_rate": 4.747781798918068e-06, "loss": 1.4962, "step": 1373 }, { "epoch": 0.9202947086403215, "grad_norm": 2.0846708164637775, "learning_rate": 4.7471417009526214e-06, "loss": 1.7245, "step": 1374 }, { "epoch": 0.9209645010046885, "grad_norm": 2.4474490431421407, "learning_rate": 4.746500835033773e-06, "loss": 1.6893, "step": 1375 }, { "epoch": 0.9216342933690556, "grad_norm": 3.6006548427952874, "learning_rate": 4.745859201380539e-06, "loss": 1.6424, "step": 1376 }, { "epoch": 0.9223040857334226, "grad_norm": 3.6217678917047866, "learning_rate": 4.7452168002121935e-06, "loss": 1.4687, "step": 1377 }, { "epoch": 0.9229738780977896, "grad_norm": 2.2893965678837436, "learning_rate": 4.744573631748276e-06, "loss": 1.6566, "step": 1378 }, { "epoch": 0.9236436704621568, "grad_norm": 2.2326378786991765, "learning_rate": 4.743929696208587e-06, "loss": 1.5926, "step": 1379 }, { "epoch": 0.9243134628265238, "grad_norm": 2.6416172234586597, "learning_rate": 4.743284993813189e-06, "loss": 1.6175, "step": 1380 }, { "epoch": 0.9249832551908909, "grad_norm": 2.8077389271531046, "learning_rate": 4.742639524782408e-06, "loss": 1.4904, "step": 1381 }, { "epoch": 0.9256530475552579, "grad_norm": 5.6563971997328295, "learning_rate": 4.74199328933683e-06, "loss": 1.6243, "step": 1382 }, { "epoch": 0.9263228399196249, "grad_norm": 3.1276395677292217, "learning_rate": 4.7413462876973035e-06, "loss": 1.549, "step": 1383 }, { "epoch": 0.926992632283992, "grad_norm": 3.0602614749514028, "learning_rate": 4.740698520084941e-06, "loss": 1.6966, "step": 1384 }, { "epoch": 0.927662424648359, "grad_norm": 2.3937396874410584, "learning_rate": 4.740049986721112e-06, "loss": 1.6614, "step": 1385 }, { "epoch": 0.9283322170127261, "grad_norm": 2.853183907790558, "learning_rate": 4.739400687827454e-06, "loss": 1.6406, "step": 1386 }, { "epoch": 0.9290020093770931, "grad_norm": 2.1649845288784233, "learning_rate": 4.738750623625862e-06, "loss": 1.5912, "step": 1387 }, { "epoch": 0.9296718017414601, "grad_norm": 3.4469254313475983, "learning_rate": 4.7380997943384925e-06, "loss": 1.9351, "step": 1388 }, { "epoch": 0.9303415941058272, "grad_norm": 2.3900273510740964, "learning_rate": 4.737448200187763e-06, "loss": 1.6555, "step": 1389 }, { "epoch": 0.9310113864701942, "grad_norm": 3.792769457235941, "learning_rate": 4.736795841396356e-06, "loss": 1.4496, "step": 1390 }, { "epoch": 0.9316811788345613, "grad_norm": 2.12491864487864, "learning_rate": 4.7361427181872125e-06, "loss": 1.7149, "step": 1391 }, { "epoch": 0.9323509711989283, "grad_norm": 2.1786872818285685, "learning_rate": 4.7354888307835344e-06, "loss": 1.5915, "step": 1392 }, { "epoch": 0.9330207635632953, "grad_norm": 2.1199294808102684, "learning_rate": 4.734834179408786e-06, "loss": 1.4887, "step": 1393 }, { "epoch": 0.9336905559276625, "grad_norm": 2.2801244634455635, "learning_rate": 4.7341787642866935e-06, "loss": 1.7185, "step": 1394 }, { "epoch": 0.9343603482920295, "grad_norm": 2.2971544968558275, "learning_rate": 4.73352258564124e-06, "loss": 1.4553, "step": 1395 }, { "epoch": 0.9350301406563966, "grad_norm": 2.25417997313609, "learning_rate": 4.732865643696676e-06, "loss": 1.6583, "step": 1396 }, { "epoch": 0.9356999330207636, "grad_norm": 3.133335817874879, "learning_rate": 4.732207938677507e-06, "loss": 1.3482, "step": 1397 }, { "epoch": 0.9363697253851306, "grad_norm": 3.3734337492435764, "learning_rate": 4.731549470808501e-06, "loss": 1.6912, "step": 1398 }, { "epoch": 0.9370395177494977, "grad_norm": 3.217222256062629, "learning_rate": 4.73089024031469e-06, "loss": 1.5712, "step": 1399 }, { "epoch": 0.9377093101138647, "grad_norm": 2.208224823853331, "learning_rate": 4.730230247421361e-06, "loss": 1.7045, "step": 1400 }, { "epoch": 0.9383791024782318, "grad_norm": 2.2091205754167134, "learning_rate": 4.729569492354066e-06, "loss": 1.6304, "step": 1401 }, { "epoch": 0.9390488948425988, "grad_norm": 3.2331399264359177, "learning_rate": 4.728907975338615e-06, "loss": 1.6253, "step": 1402 }, { "epoch": 0.9397186872069658, "grad_norm": 2.2750745860077055, "learning_rate": 4.728245696601081e-06, "loss": 1.6738, "step": 1403 }, { "epoch": 0.9403884795713329, "grad_norm": 6.846078251182177, "learning_rate": 4.727582656367794e-06, "loss": 1.4388, "step": 1404 }, { "epoch": 0.9410582719356999, "grad_norm": 2.315739707530677, "learning_rate": 4.726918854865345e-06, "loss": 1.5396, "step": 1405 }, { "epoch": 0.941728064300067, "grad_norm": 2.2287640215076983, "learning_rate": 4.726254292320589e-06, "loss": 1.6987, "step": 1406 }, { "epoch": 0.942397856664434, "grad_norm": 2.4105214245656663, "learning_rate": 4.725588968960636e-06, "loss": 1.7153, "step": 1407 }, { "epoch": 0.943067649028801, "grad_norm": 7.716908749302143, "learning_rate": 4.724922885012859e-06, "loss": 1.6619, "step": 1408 }, { "epoch": 0.9437374413931681, "grad_norm": 3.396063974318238, "learning_rate": 4.7242560407048886e-06, "loss": 1.7735, "step": 1409 }, { "epoch": 0.9444072337575352, "grad_norm": 2.91022681182736, "learning_rate": 4.723588436264619e-06, "loss": 1.5578, "step": 1410 }, { "epoch": 0.9450770261219023, "grad_norm": 2.5024375084869237, "learning_rate": 4.7229200719201994e-06, "loss": 1.6073, "step": 1411 }, { "epoch": 0.9457468184862693, "grad_norm": 2.3690297264799582, "learning_rate": 4.7222509479000425e-06, "loss": 1.7955, "step": 1412 }, { "epoch": 0.9464166108506363, "grad_norm": 2.1886930761220404, "learning_rate": 4.721581064432818e-06, "loss": 1.6044, "step": 1413 }, { "epoch": 0.9470864032150034, "grad_norm": 6.40149800100863, "learning_rate": 4.7209104217474595e-06, "loss": 1.8359, "step": 1414 }, { "epoch": 0.9477561955793704, "grad_norm": 5.725761240718178, "learning_rate": 4.720239020073154e-06, "loss": 1.7627, "step": 1415 }, { "epoch": 0.9484259879437374, "grad_norm": 2.1743541123200685, "learning_rate": 4.719566859639352e-06, "loss": 1.5848, "step": 1416 }, { "epoch": 0.9490957803081045, "grad_norm": 2.2778638593262106, "learning_rate": 4.7188939406757615e-06, "loss": 1.747, "step": 1417 }, { "epoch": 0.9497655726724715, "grad_norm": 5.7663885629774825, "learning_rate": 4.71822026341235e-06, "loss": 1.8435, "step": 1418 }, { "epoch": 0.9504353650368386, "grad_norm": 3.8814814170222314, "learning_rate": 4.717545828079347e-06, "loss": 1.6643, "step": 1419 }, { "epoch": 0.9511051574012056, "grad_norm": 2.25327645369436, "learning_rate": 4.7168706349072355e-06, "loss": 1.8081, "step": 1420 }, { "epoch": 0.9517749497655726, "grad_norm": 3.257224646631471, "learning_rate": 4.716194684126762e-06, "loss": 1.7733, "step": 1421 }, { "epoch": 0.9524447421299397, "grad_norm": 2.3977455031573003, "learning_rate": 4.715517975968932e-06, "loss": 1.7018, "step": 1422 }, { "epoch": 0.9531145344943067, "grad_norm": 2.622503611810477, "learning_rate": 4.7148405106650055e-06, "loss": 1.6502, "step": 1423 }, { "epoch": 0.9537843268586738, "grad_norm": 3.9007871056960264, "learning_rate": 4.714162288446505e-06, "loss": 1.709, "step": 1424 }, { "epoch": 0.9544541192230409, "grad_norm": 2.260071313651845, "learning_rate": 4.713483309545212e-06, "loss": 1.7835, "step": 1425 }, { "epoch": 0.955123911587408, "grad_norm": 2.1800063226685897, "learning_rate": 4.712803574193165e-06, "loss": 1.6891, "step": 1426 }, { "epoch": 0.955793703951775, "grad_norm": 2.3544298850463634, "learning_rate": 4.71212308262266e-06, "loss": 1.6616, "step": 1427 }, { "epoch": 0.956463496316142, "grad_norm": 2.9784239658641978, "learning_rate": 4.711441835066253e-06, "loss": 1.5257, "step": 1428 }, { "epoch": 0.9571332886805091, "grad_norm": 3.0998195935398503, "learning_rate": 4.710759831756759e-06, "loss": 1.6393, "step": 1429 }, { "epoch": 0.9578030810448761, "grad_norm": 2.5223955693914193, "learning_rate": 4.71007707292725e-06, "loss": 1.3981, "step": 1430 }, { "epoch": 0.9584728734092431, "grad_norm": 48.63871950638264, "learning_rate": 4.7093935588110585e-06, "loss": 2.092, "step": 1431 }, { "epoch": 0.9591426657736102, "grad_norm": 2.9880864575193553, "learning_rate": 4.708709289641769e-06, "loss": 1.7948, "step": 1432 }, { "epoch": 0.9598124581379772, "grad_norm": 5.0146142340974675, "learning_rate": 4.7080242656532326e-06, "loss": 1.7362, "step": 1433 }, { "epoch": 0.9604822505023443, "grad_norm": 3.402970560764185, "learning_rate": 4.707338487079553e-06, "loss": 1.5467, "step": 1434 }, { "epoch": 0.9611520428667113, "grad_norm": 2.5852153447764854, "learning_rate": 4.70665195415509e-06, "loss": 1.7553, "step": 1435 }, { "epoch": 0.9618218352310783, "grad_norm": 2.2494826943619435, "learning_rate": 4.7059646671144665e-06, "loss": 1.6523, "step": 1436 }, { "epoch": 0.9624916275954454, "grad_norm": 3.432077915380634, "learning_rate": 4.705276626192561e-06, "loss": 1.5295, "step": 1437 }, { "epoch": 0.9631614199598124, "grad_norm": 3.4413498551392876, "learning_rate": 4.7045878316245075e-06, "loss": 1.5123, "step": 1438 }, { "epoch": 0.9638312123241795, "grad_norm": 8.420235074718041, "learning_rate": 4.703898283645701e-06, "loss": 1.5563, "step": 1439 }, { "epoch": 0.9645010046885466, "grad_norm": 2.831869357999571, "learning_rate": 4.703207982491791e-06, "loss": 1.5442, "step": 1440 }, { "epoch": 0.9651707970529136, "grad_norm": 2.342809909168546, "learning_rate": 4.702516928398686e-06, "loss": 1.7643, "step": 1441 }, { "epoch": 0.9658405894172807, "grad_norm": 2.3611867357714718, "learning_rate": 4.7018251216025525e-06, "loss": 1.5518, "step": 1442 }, { "epoch": 0.9665103817816477, "grad_norm": 2.175436572716867, "learning_rate": 4.701132562339812e-06, "loss": 1.5264, "step": 1443 }, { "epoch": 0.9671801741460148, "grad_norm": 2.2265187525429675, "learning_rate": 4.7004392508471445e-06, "loss": 1.6481, "step": 1444 }, { "epoch": 0.9678499665103818, "grad_norm": 3.5741576328605036, "learning_rate": 4.699745187361487e-06, "loss": 1.7555, "step": 1445 }, { "epoch": 0.9685197588747488, "grad_norm": 3.629207747070085, "learning_rate": 4.699050372120035e-06, "loss": 1.5251, "step": 1446 }, { "epoch": 0.9691895512391159, "grad_norm": 4.388647662378873, "learning_rate": 4.698354805360237e-06, "loss": 1.4772, "step": 1447 }, { "epoch": 0.9698593436034829, "grad_norm": 4.082775848495771, "learning_rate": 4.697658487319803e-06, "loss": 1.7136, "step": 1448 }, { "epoch": 0.97052913596785, "grad_norm": 3.219588332926164, "learning_rate": 4.696961418236695e-06, "loss": 1.6684, "step": 1449 }, { "epoch": 0.971198928332217, "grad_norm": 2.1522673008346542, "learning_rate": 4.696263598349136e-06, "loss": 1.5883, "step": 1450 }, { "epoch": 0.971868720696584, "grad_norm": 2.751748639889232, "learning_rate": 4.6955650278956025e-06, "loss": 1.7756, "step": 1451 }, { "epoch": 0.9725385130609511, "grad_norm": 3.9514759030836997, "learning_rate": 4.69486570711483e-06, "loss": 1.7261, "step": 1452 }, { "epoch": 0.9732083054253181, "grad_norm": 2.931381145100206, "learning_rate": 4.6941656362458074e-06, "loss": 1.7571, "step": 1453 }, { "epoch": 0.9738780977896851, "grad_norm": 3.4068632607977567, "learning_rate": 4.693464815527783e-06, "loss": 1.6159, "step": 1454 }, { "epoch": 0.9745478901540523, "grad_norm": 2.311801584563611, "learning_rate": 4.6927632452002595e-06, "loss": 1.7398, "step": 1455 }, { "epoch": 0.9752176825184193, "grad_norm": 2.278203758128678, "learning_rate": 4.692060925502996e-06, "loss": 1.623, "step": 1456 }, { "epoch": 0.9758874748827864, "grad_norm": 2.2992592797233704, "learning_rate": 4.691357856676009e-06, "loss": 1.7454, "step": 1457 }, { "epoch": 0.9765572672471534, "grad_norm": 2.746224510251606, "learning_rate": 4.690654038959568e-06, "loss": 1.7584, "step": 1458 }, { "epoch": 0.9772270596115205, "grad_norm": 2.7631675738016805, "learning_rate": 4.689949472594203e-06, "loss": 1.5271, "step": 1459 }, { "epoch": 0.9778968519758875, "grad_norm": 2.2995774842163397, "learning_rate": 4.689244157820696e-06, "loss": 1.4978, "step": 1460 }, { "epoch": 0.9785666443402545, "grad_norm": 3.9846483016155143, "learning_rate": 4.6885380948800854e-06, "loss": 1.5808, "step": 1461 }, { "epoch": 0.9792364367046216, "grad_norm": 3.372916648517565, "learning_rate": 4.6878312840136675e-06, "loss": 1.6249, "step": 1462 }, { "epoch": 0.9799062290689886, "grad_norm": 3.515321110544292, "learning_rate": 4.687123725462991e-06, "loss": 1.654, "step": 1463 }, { "epoch": 0.9805760214333556, "grad_norm": 2.775205703660844, "learning_rate": 4.686415419469863e-06, "loss": 1.4802, "step": 1464 }, { "epoch": 0.9812458137977227, "grad_norm": 2.3773091685665912, "learning_rate": 4.685706366276344e-06, "loss": 1.6745, "step": 1465 }, { "epoch": 0.9819156061620897, "grad_norm": 4.079121578156953, "learning_rate": 4.684996566124751e-06, "loss": 1.6097, "step": 1466 }, { "epoch": 0.9825853985264568, "grad_norm": 2.463678972034602, "learning_rate": 4.684286019257656e-06, "loss": 1.5873, "step": 1467 }, { "epoch": 0.9832551908908238, "grad_norm": 7.087815487982806, "learning_rate": 4.683574725917884e-06, "loss": 1.6639, "step": 1468 }, { "epoch": 0.9839249832551908, "grad_norm": 4.086019182996725, "learning_rate": 4.682862686348522e-06, "loss": 1.6366, "step": 1469 }, { "epoch": 0.9845947756195579, "grad_norm": 2.275545939517765, "learning_rate": 4.682149900792902e-06, "loss": 1.6899, "step": 1470 }, { "epoch": 0.985264567983925, "grad_norm": 3.5559243119902413, "learning_rate": 4.68143636949462e-06, "loss": 1.566, "step": 1471 }, { "epoch": 0.9859343603482921, "grad_norm": 2.5585904329552154, "learning_rate": 4.680722092697521e-06, "loss": 1.5675, "step": 1472 }, { "epoch": 0.9866041527126591, "grad_norm": 2.2434994360887632, "learning_rate": 4.6800070706457065e-06, "loss": 1.6081, "step": 1473 }, { "epoch": 0.9872739450770261, "grad_norm": 3.091307386463921, "learning_rate": 4.679291303583534e-06, "loss": 1.7472, "step": 1474 }, { "epoch": 0.9879437374413932, "grad_norm": 2.1939572479143306, "learning_rate": 4.6785747917556145e-06, "loss": 1.716, "step": 1475 }, { "epoch": 0.9886135298057602, "grad_norm": 2.957312703636323, "learning_rate": 4.677857535406813e-06, "loss": 1.4814, "step": 1476 }, { "epoch": 0.9892833221701273, "grad_norm": 2.268422006747675, "learning_rate": 4.67713953478225e-06, "loss": 1.6636, "step": 1477 }, { "epoch": 0.9899531145344943, "grad_norm": 2.9250251244625716, "learning_rate": 4.676420790127301e-06, "loss": 1.7375, "step": 1478 }, { "epoch": 0.9906229068988613, "grad_norm": 3.191986914763822, "learning_rate": 4.675701301687592e-06, "loss": 1.5896, "step": 1479 }, { "epoch": 0.9912926992632284, "grad_norm": 2.1977760349933426, "learning_rate": 4.674981069709008e-06, "loss": 1.6554, "step": 1480 }, { "epoch": 0.9919624916275954, "grad_norm": 2.864354587839654, "learning_rate": 4.674260094437685e-06, "loss": 1.8376, "step": 1481 }, { "epoch": 0.9926322839919625, "grad_norm": 5.094178744005397, "learning_rate": 4.673538376120015e-06, "loss": 1.3459, "step": 1482 }, { "epoch": 0.9933020763563295, "grad_norm": 2.2990893920691, "learning_rate": 4.672815915002642e-06, "loss": 1.7245, "step": 1483 }, { "epoch": 0.9939718687206965, "grad_norm": 14.124928524577257, "learning_rate": 4.672092711332464e-06, "loss": 2.001, "step": 1484 }, { "epoch": 0.9946416610850636, "grad_norm": 2.814795970498044, "learning_rate": 4.671368765356634e-06, "loss": 1.7131, "step": 1485 }, { "epoch": 0.9953114534494307, "grad_norm": 2.241101463133091, "learning_rate": 4.67064407732256e-06, "loss": 1.3851, "step": 1486 }, { "epoch": 0.9959812458137978, "grad_norm": 2.1623496901855654, "learning_rate": 4.6699186474779e-06, "loss": 1.6318, "step": 1487 }, { "epoch": 0.9966510381781648, "grad_norm": 2.2537599804049195, "learning_rate": 4.669192476070567e-06, "loss": 1.6207, "step": 1488 }, { "epoch": 0.9973208305425318, "grad_norm": 2.2007840088546042, "learning_rate": 4.6684655633487285e-06, "loss": 1.8128, "step": 1489 }, { "epoch": 0.9979906229068989, "grad_norm": 2.95630214502749, "learning_rate": 4.667737909560805e-06, "loss": 1.5764, "step": 1490 }, { "epoch": 0.9986604152712659, "grad_norm": 3.214263911810492, "learning_rate": 4.667009514955469e-06, "loss": 1.664, "step": 1491 }, { "epoch": 0.999330207635633, "grad_norm": 6.86915896346729, "learning_rate": 4.666280379781647e-06, "loss": 1.5571, "step": 1492 }, { "epoch": 1.0, "grad_norm": 2.2984781436147603, "learning_rate": 4.66555050428852e-06, "loss": 1.6665, "step": 1493 }, { "epoch": 1.000669792364367, "grad_norm": 2.463270616457345, "learning_rate": 4.664819888725518e-06, "loss": 1.7036, "step": 1494 }, { "epoch": 1.001339584728734, "grad_norm": 2.444981590503454, "learning_rate": 4.664088533342329e-06, "loss": 1.5524, "step": 1495 }, { "epoch": 1.0020093770931011, "grad_norm": 49.29389324568308, "learning_rate": 4.66335643838889e-06, "loss": 1.7353, "step": 1496 }, { "epoch": 1.0026791694574682, "grad_norm": 2.2612398926529846, "learning_rate": 4.662623604115392e-06, "loss": 1.4858, "step": 1497 }, { "epoch": 1.0033489618218352, "grad_norm": 2.2117150042897284, "learning_rate": 4.66189003077228e-06, "loss": 1.466, "step": 1498 }, { "epoch": 1.0040187541862022, "grad_norm": 2.592491373046008, "learning_rate": 4.661155718610248e-06, "loss": 1.6365, "step": 1499 }, { "epoch": 1.0046885465505693, "grad_norm": 2.5066195123197827, "learning_rate": 4.660420667880247e-06, "loss": 1.5599, "step": 1500 }, { "epoch": 1.0053583389149363, "grad_norm": 2.325249270570369, "learning_rate": 4.659684878833478e-06, "loss": 1.5531, "step": 1501 }, { "epoch": 1.0060281312793034, "grad_norm": 2.410707289387868, "learning_rate": 4.658948351721393e-06, "loss": 1.665, "step": 1502 }, { "epoch": 1.0066979236436704, "grad_norm": 97.60028437402156, "learning_rate": 4.658211086795699e-06, "loss": 1.5401, "step": 1503 }, { "epoch": 1.0073677160080374, "grad_norm": 2.462611088756253, "learning_rate": 4.657473084308354e-06, "loss": 1.4435, "step": 1504 }, { "epoch": 1.0080375083724045, "grad_norm": 2.5951525577159624, "learning_rate": 4.656734344511568e-06, "loss": 1.6128, "step": 1505 }, { "epoch": 1.0087073007367715, "grad_norm": 2.1945533046151144, "learning_rate": 4.655994867657802e-06, "loss": 1.4541, "step": 1506 }, { "epoch": 1.0093770931011385, "grad_norm": 3.7314983698954665, "learning_rate": 4.65525465399977e-06, "loss": 1.3616, "step": 1507 }, { "epoch": 1.0100468854655056, "grad_norm": 2.398574602836736, "learning_rate": 4.654513703790438e-06, "loss": 1.421, "step": 1508 }, { "epoch": 1.0107166778298728, "grad_norm": 2.4590174098444173, "learning_rate": 4.653772017283024e-06, "loss": 1.6202, "step": 1509 }, { "epoch": 1.0113864701942399, "grad_norm": 2.408537425743249, "learning_rate": 4.653029594730995e-06, "loss": 1.5212, "step": 1510 }, { "epoch": 1.012056262558607, "grad_norm": 2.8720773785259786, "learning_rate": 4.652286436388074e-06, "loss": 1.5139, "step": 1511 }, { "epoch": 1.012726054922974, "grad_norm": 2.5224473258896585, "learning_rate": 4.651542542508231e-06, "loss": 1.1965, "step": 1512 }, { "epoch": 1.013395847287341, "grad_norm": 2.3874456822130035, "learning_rate": 4.65079791334569e-06, "loss": 1.5821, "step": 1513 }, { "epoch": 1.014065639651708, "grad_norm": 2.385227786581948, "learning_rate": 4.650052549154926e-06, "loss": 1.5994, "step": 1514 }, { "epoch": 1.014735432016075, "grad_norm": 4.0924455552504195, "learning_rate": 4.6493064501906646e-06, "loss": 1.3805, "step": 1515 }, { "epoch": 1.0154052243804421, "grad_norm": 2.235355955712157, "learning_rate": 4.648559616707881e-06, "loss": 1.5603, "step": 1516 }, { "epoch": 1.0160750167448092, "grad_norm": 3.959317948987152, "learning_rate": 4.647812048961807e-06, "loss": 1.6229, "step": 1517 }, { "epoch": 1.0167448091091762, "grad_norm": 2.3327605060522263, "learning_rate": 4.6470637472079185e-06, "loss": 1.4919, "step": 1518 }, { "epoch": 1.0174146014735432, "grad_norm": 3.310080990995966, "learning_rate": 4.646314711701945e-06, "loss": 1.5755, "step": 1519 }, { "epoch": 1.0180843938379103, "grad_norm": 2.2520745714143344, "learning_rate": 4.645564942699869e-06, "loss": 1.4859, "step": 1520 }, { "epoch": 1.0187541862022773, "grad_norm": 2.568608714538228, "learning_rate": 4.6448144404579186e-06, "loss": 1.7422, "step": 1521 }, { "epoch": 1.0194239785666444, "grad_norm": 3.69468661249459, "learning_rate": 4.644063205232579e-06, "loss": 1.4088, "step": 1522 }, { "epoch": 1.0200937709310114, "grad_norm": 2.218929547305462, "learning_rate": 4.64331123728058e-06, "loss": 1.4397, "step": 1523 }, { "epoch": 1.0207635632953784, "grad_norm": 2.313837487855527, "learning_rate": 4.6425585368589056e-06, "loss": 1.5108, "step": 1524 }, { "epoch": 1.0214333556597455, "grad_norm": 2.357877981832022, "learning_rate": 4.6418051042247865e-06, "loss": 1.4404, "step": 1525 }, { "epoch": 1.0221031480241125, "grad_norm": 3.1321984922088135, "learning_rate": 4.64105093963571e-06, "loss": 1.6696, "step": 1526 }, { "epoch": 1.0227729403884795, "grad_norm": 2.3199553059474898, "learning_rate": 4.640296043349405e-06, "loss": 1.5011, "step": 1527 }, { "epoch": 1.0234427327528466, "grad_norm": 2.768994648493535, "learning_rate": 4.639540415623857e-06, "loss": 1.6448, "step": 1528 }, { "epoch": 1.0241125251172136, "grad_norm": 2.353646555710916, "learning_rate": 4.638784056717299e-06, "loss": 1.513, "step": 1529 }, { "epoch": 1.0247823174815807, "grad_norm": 2.375839512506609, "learning_rate": 4.638026966888214e-06, "loss": 1.4276, "step": 1530 }, { "epoch": 1.0254521098459477, "grad_norm": 4.344666428487511, "learning_rate": 4.637269146395336e-06, "loss": 1.4362, "step": 1531 }, { "epoch": 1.0261219022103147, "grad_norm": 3.0889827549183533, "learning_rate": 4.636510595497647e-06, "loss": 1.428, "step": 1532 }, { "epoch": 1.0267916945746818, "grad_norm": 2.9317422501676926, "learning_rate": 4.635751314454379e-06, "loss": 1.5187, "step": 1533 }, { "epoch": 1.0274614869390488, "grad_norm": 3.601660059847983, "learning_rate": 4.634991303525014e-06, "loss": 1.4456, "step": 1534 }, { "epoch": 1.0281312793034159, "grad_norm": 4.797792563761114, "learning_rate": 4.634230562969283e-06, "loss": 1.6274, "step": 1535 }, { "epoch": 1.028801071667783, "grad_norm": 3.3337078409455527, "learning_rate": 4.633469093047167e-06, "loss": 1.2479, "step": 1536 }, { "epoch": 1.02947086403215, "grad_norm": 6.282977909389164, "learning_rate": 4.632706894018897e-06, "loss": 1.4127, "step": 1537 }, { "epoch": 1.030140656396517, "grad_norm": 3.6195186292000314, "learning_rate": 4.63194396614495e-06, "loss": 1.4033, "step": 1538 }, { "epoch": 1.0308104487608842, "grad_norm": 2.680073110886144, "learning_rate": 4.6311803096860555e-06, "loss": 1.5709, "step": 1539 }, { "epoch": 1.0314802411252513, "grad_norm": 2.6033439794130735, "learning_rate": 4.63041592490319e-06, "loss": 1.3974, "step": 1540 }, { "epoch": 1.0321500334896183, "grad_norm": 2.3984570101941634, "learning_rate": 4.62965081205758e-06, "loss": 1.4601, "step": 1541 }, { "epoch": 1.0328198258539854, "grad_norm": 2.474125592319297, "learning_rate": 4.6288849714107e-06, "loss": 1.4697, "step": 1542 }, { "epoch": 1.0334896182183524, "grad_norm": 2.389646749188778, "learning_rate": 4.628118403224273e-06, "loss": 1.6213, "step": 1543 }, { "epoch": 1.0341594105827194, "grad_norm": 2.2072605717739724, "learning_rate": 4.627351107760272e-06, "loss": 1.6393, "step": 1544 }, { "epoch": 1.0348292029470865, "grad_norm": 2.509193367685455, "learning_rate": 4.626583085280918e-06, "loss": 1.4293, "step": 1545 }, { "epoch": 1.0354989953114535, "grad_norm": 2.22896148479444, "learning_rate": 4.625814336048679e-06, "loss": 1.5681, "step": 1546 }, { "epoch": 1.0361687876758205, "grad_norm": 2.274542208738713, "learning_rate": 4.625044860326274e-06, "loss": 1.4651, "step": 1547 }, { "epoch": 1.0368385800401876, "grad_norm": 3.7581313378625643, "learning_rate": 4.624274658376667e-06, "loss": 1.3497, "step": 1548 }, { "epoch": 1.0375083724045546, "grad_norm": 3.0327116864286885, "learning_rate": 4.623503730463075e-06, "loss": 1.4129, "step": 1549 }, { "epoch": 1.0381781647689217, "grad_norm": 2.3883565253541676, "learning_rate": 4.622732076848957e-06, "loss": 1.4242, "step": 1550 }, { "epoch": 1.0388479571332887, "grad_norm": 2.366072374722946, "learning_rate": 4.621959697798024e-06, "loss": 1.4827, "step": 1551 }, { "epoch": 1.0395177494976557, "grad_norm": 4.720631683379491, "learning_rate": 4.621186593574236e-06, "loss": 1.3894, "step": 1552 }, { "epoch": 1.0401875418620228, "grad_norm": 2.592494078830646, "learning_rate": 4.620412764441796e-06, "loss": 1.5084, "step": 1553 }, { "epoch": 1.0408573342263898, "grad_norm": 2.3228112938960654, "learning_rate": 4.619638210665159e-06, "loss": 1.4551, "step": 1554 }, { "epoch": 1.0415271265907569, "grad_norm": 2.4123222945303295, "learning_rate": 4.6188629325090276e-06, "loss": 1.7166, "step": 1555 }, { "epoch": 1.042196918955124, "grad_norm": 3.328678284437913, "learning_rate": 4.618086930238348e-06, "loss": 1.4769, "step": 1556 }, { "epoch": 1.042866711319491, "grad_norm": 2.81483051397274, "learning_rate": 4.617310204118317e-06, "loss": 1.4489, "step": 1557 }, { "epoch": 1.043536503683858, "grad_norm": 2.371771715331195, "learning_rate": 4.61653275441438e-06, "loss": 1.5183, "step": 1558 }, { "epoch": 1.044206296048225, "grad_norm": 3.1241449379488593, "learning_rate": 4.6157545813922265e-06, "loss": 1.4501, "step": 1559 }, { "epoch": 1.044876088412592, "grad_norm": 3.1279088222213978, "learning_rate": 4.614975685317794e-06, "loss": 1.2935, "step": 1560 }, { "epoch": 1.045545880776959, "grad_norm": 3.9766883114349363, "learning_rate": 4.61419606645727e-06, "loss": 1.3234, "step": 1561 }, { "epoch": 1.0462156731413261, "grad_norm": 3.6735949196295765, "learning_rate": 4.613415725077084e-06, "loss": 1.3475, "step": 1562 }, { "epoch": 1.0468854655056932, "grad_norm": 2.270273149349959, "learning_rate": 4.612634661443918e-06, "loss": 1.4246, "step": 1563 }, { "epoch": 1.0475552578700602, "grad_norm": 2.3898876051591476, "learning_rate": 4.611852875824695e-06, "loss": 1.6636, "step": 1564 }, { "epoch": 1.0482250502344272, "grad_norm": 5.360999894258614, "learning_rate": 4.61107036848659e-06, "loss": 1.4705, "step": 1565 }, { "epoch": 1.0488948425987943, "grad_norm": 3.9046108057980216, "learning_rate": 4.610287139697021e-06, "loss": 1.5711, "step": 1566 }, { "epoch": 1.0495646349631613, "grad_norm": 2.8611160948116265, "learning_rate": 4.609503189723654e-06, "loss": 1.4037, "step": 1567 }, { "epoch": 1.0502344273275284, "grad_norm": 4.776905134729466, "learning_rate": 4.608718518834402e-06, "loss": 1.4624, "step": 1568 }, { "epoch": 1.0509042196918954, "grad_norm": 2.452719201625053, "learning_rate": 4.6079331272974246e-06, "loss": 1.4918, "step": 1569 }, { "epoch": 1.0515740120562627, "grad_norm": 2.2560571760647887, "learning_rate": 4.6071470153811245e-06, "loss": 1.4013, "step": 1570 }, { "epoch": 1.0522438044206297, "grad_norm": 5.875971426652392, "learning_rate": 4.606360183354155e-06, "loss": 1.2504, "step": 1571 }, { "epoch": 1.0529135967849967, "grad_norm": 2.148657512063248, "learning_rate": 4.605572631485412e-06, "loss": 1.4263, "step": 1572 }, { "epoch": 1.0535833891493638, "grad_norm": 2.2859289474132303, "learning_rate": 4.604784360044039e-06, "loss": 1.5224, "step": 1573 }, { "epoch": 1.0542531815137308, "grad_norm": 3.4462472304820873, "learning_rate": 4.603995369299425e-06, "loss": 1.5502, "step": 1574 }, { "epoch": 1.0549229738780979, "grad_norm": 2.2353950136053164, "learning_rate": 4.603205659521207e-06, "loss": 1.2953, "step": 1575 }, { "epoch": 1.055592766242465, "grad_norm": 2.421356080903517, "learning_rate": 4.602415230979264e-06, "loss": 1.6569, "step": 1576 }, { "epoch": 1.056262558606832, "grad_norm": 2.3069835035961623, "learning_rate": 4.601624083943722e-06, "loss": 1.4048, "step": 1577 }, { "epoch": 1.056932350971199, "grad_norm": 2.8502133305156687, "learning_rate": 4.600832218684954e-06, "loss": 1.4176, "step": 1578 }, { "epoch": 1.057602143335566, "grad_norm": 2.0925940425088787, "learning_rate": 4.600039635473577e-06, "loss": 1.3683, "step": 1579 }, { "epoch": 1.058271935699933, "grad_norm": 2.262157062569138, "learning_rate": 4.5992463345804536e-06, "loss": 1.4784, "step": 1580 }, { "epoch": 1.0589417280643, "grad_norm": 2.1301285540041452, "learning_rate": 4.598452316276692e-06, "loss": 1.3281, "step": 1581 }, { "epoch": 1.0596115204286671, "grad_norm": 2.2771571415490275, "learning_rate": 4.5976575808336455e-06, "loss": 1.4576, "step": 1582 }, { "epoch": 1.0602813127930342, "grad_norm": 2.236055817353062, "learning_rate": 4.596862128522912e-06, "loss": 1.2666, "step": 1583 }, { "epoch": 1.0609511051574012, "grad_norm": 3.667283750750661, "learning_rate": 4.596065959616336e-06, "loss": 1.5342, "step": 1584 }, { "epoch": 1.0616208975217682, "grad_norm": 2.3975726993719984, "learning_rate": 4.595269074386003e-06, "loss": 1.5388, "step": 1585 }, { "epoch": 1.0622906898861353, "grad_norm": 2.269895779830727, "learning_rate": 4.5944714731042486e-06, "loss": 1.555, "step": 1586 }, { "epoch": 1.0629604822505023, "grad_norm": 2.3620226339617063, "learning_rate": 4.5936731560436495e-06, "loss": 1.3346, "step": 1587 }, { "epoch": 1.0636302746148694, "grad_norm": 2.5211306172492147, "learning_rate": 4.592874123477028e-06, "loss": 1.2606, "step": 1588 }, { "epoch": 1.0643000669792364, "grad_norm": 3.1700420378166525, "learning_rate": 4.5920743756774505e-06, "loss": 1.4178, "step": 1589 }, { "epoch": 1.0649698593436034, "grad_norm": 3.21179410593773, "learning_rate": 4.591273912918228e-06, "loss": 1.5301, "step": 1590 }, { "epoch": 1.0656396517079705, "grad_norm": 2.484866496515579, "learning_rate": 4.590472735472917e-06, "loss": 1.59, "step": 1591 }, { "epoch": 1.0663094440723375, "grad_norm": 2.401250475092848, "learning_rate": 4.5896708436153174e-06, "loss": 1.4471, "step": 1592 }, { "epoch": 1.0669792364367046, "grad_norm": 2.231992864262664, "learning_rate": 4.5888682376194726e-06, "loss": 1.5345, "step": 1593 }, { "epoch": 1.0676490288010716, "grad_norm": 2.2945115296536485, "learning_rate": 4.588064917759671e-06, "loss": 1.4165, "step": 1594 }, { "epoch": 1.0683188211654386, "grad_norm": 2.3352330672357304, "learning_rate": 4.587260884310443e-06, "loss": 1.437, "step": 1595 }, { "epoch": 1.0689886135298057, "grad_norm": 2.2735137481187295, "learning_rate": 4.5864561375465654e-06, "loss": 1.2743, "step": 1596 }, { "epoch": 1.0696584058941727, "grad_norm": 2.1833318485929762, "learning_rate": 4.5856506777430585e-06, "loss": 1.3233, "step": 1597 }, { "epoch": 1.0703281982585398, "grad_norm": 3.695213458460917, "learning_rate": 4.584844505175186e-06, "loss": 1.4979, "step": 1598 }, { "epoch": 1.070997990622907, "grad_norm": 3.1089582479905533, "learning_rate": 4.584037620118452e-06, "loss": 1.417, "step": 1599 }, { "epoch": 1.0716677829872738, "grad_norm": 3.4753968408885365, "learning_rate": 4.58323002284861e-06, "loss": 1.3359, "step": 1600 }, { "epoch": 1.072337575351641, "grad_norm": 2.86888442012518, "learning_rate": 4.582421713641653e-06, "loss": 1.4231, "step": 1601 }, { "epoch": 1.0730073677160081, "grad_norm": 2.449388317646692, "learning_rate": 4.5816126927738165e-06, "loss": 1.4423, "step": 1602 }, { "epoch": 1.0736771600803752, "grad_norm": 2.3699452383549158, "learning_rate": 4.580802960521582e-06, "loss": 1.6212, "step": 1603 }, { "epoch": 1.0743469524447422, "grad_norm": 2.8958660061832027, "learning_rate": 4.579992517161672e-06, "loss": 1.4985, "step": 1604 }, { "epoch": 1.0750167448091092, "grad_norm": 2.0133451647729688, "learning_rate": 4.579181362971055e-06, "loss": 1.2633, "step": 1605 }, { "epoch": 1.0756865371734763, "grad_norm": 2.9237241876626245, "learning_rate": 4.578369498226938e-06, "loss": 1.6174, "step": 1606 }, { "epoch": 1.0763563295378433, "grad_norm": 2.6185388031781374, "learning_rate": 4.577556923206774e-06, "loss": 1.6497, "step": 1607 }, { "epoch": 1.0770261219022104, "grad_norm": 7.708135021349874, "learning_rate": 4.576743638188258e-06, "loss": 1.5874, "step": 1608 }, { "epoch": 1.0776959142665774, "grad_norm": 2.9017632557870097, "learning_rate": 4.575929643449328e-06, "loss": 1.3684, "step": 1609 }, { "epoch": 1.0783657066309444, "grad_norm": 2.181568015055086, "learning_rate": 4.575114939268163e-06, "loss": 1.4551, "step": 1610 }, { "epoch": 1.0790354989953115, "grad_norm": 3.2786841576976005, "learning_rate": 4.574299525923186e-06, "loss": 1.2239, "step": 1611 }, { "epoch": 1.0797052913596785, "grad_norm": 2.3203844615379925, "learning_rate": 4.573483403693061e-06, "loss": 1.6202, "step": 1612 }, { "epoch": 1.0803750837240456, "grad_norm": 3.352263074016169, "learning_rate": 4.572666572856697e-06, "loss": 1.3958, "step": 1613 }, { "epoch": 1.0810448760884126, "grad_norm": 7.685023273384054, "learning_rate": 4.571849033693242e-06, "loss": 1.3527, "step": 1614 }, { "epoch": 1.0817146684527796, "grad_norm": 2.2724358229577453, "learning_rate": 4.571030786482088e-06, "loss": 1.4127, "step": 1615 }, { "epoch": 1.0823844608171467, "grad_norm": 3.5608150753282843, "learning_rate": 4.5702118315028675e-06, "loss": 1.4742, "step": 1616 }, { "epoch": 1.0830542531815137, "grad_norm": 2.9057589151021084, "learning_rate": 4.569392169035457e-06, "loss": 1.422, "step": 1617 }, { "epoch": 1.0837240455458808, "grad_norm": 4.12765663315556, "learning_rate": 4.568571799359972e-06, "loss": 1.1943, "step": 1618 }, { "epoch": 1.0843938379102478, "grad_norm": 3.0249599266379654, "learning_rate": 4.5677507227567726e-06, "loss": 1.2895, "step": 1619 }, { "epoch": 1.0850636302746148, "grad_norm": 2.769536531165139, "learning_rate": 4.566928939506457e-06, "loss": 1.6013, "step": 1620 }, { "epoch": 1.0857334226389819, "grad_norm": 2.364356537405433, "learning_rate": 4.56610644988987e-06, "loss": 1.5741, "step": 1621 }, { "epoch": 1.086403215003349, "grad_norm": 2.442933295721994, "learning_rate": 4.565283254188092e-06, "loss": 1.6378, "step": 1622 }, { "epoch": 1.087073007367716, "grad_norm": 3.6139019328514257, "learning_rate": 4.5644593526824485e-06, "loss": 1.5385, "step": 1623 }, { "epoch": 1.087742799732083, "grad_norm": 2.4193464790003265, "learning_rate": 4.563634745654506e-06, "loss": 1.5816, "step": 1624 }, { "epoch": 1.08841259209645, "grad_norm": 2.5449687135216292, "learning_rate": 4.5628094333860696e-06, "loss": 1.7578, "step": 1625 }, { "epoch": 1.089082384460817, "grad_norm": 2.8489180928894924, "learning_rate": 4.5619834161591885e-06, "loss": 1.3523, "step": 1626 }, { "epoch": 1.089752176825184, "grad_norm": 2.2950829967338646, "learning_rate": 4.5611566942561515e-06, "loss": 1.5006, "step": 1627 }, { "epoch": 1.0904219691895511, "grad_norm": 2.2004098123575755, "learning_rate": 4.5603292679594865e-06, "loss": 1.5424, "step": 1628 }, { "epoch": 1.0910917615539182, "grad_norm": 2.5683228801687936, "learning_rate": 4.559501137551965e-06, "loss": 1.6651, "step": 1629 }, { "epoch": 1.0917615539182854, "grad_norm": 2.2895321212217, "learning_rate": 4.558672303316598e-06, "loss": 1.6062, "step": 1630 }, { "epoch": 1.0924313462826523, "grad_norm": 2.6695346052342273, "learning_rate": 4.557842765536637e-06, "loss": 1.5302, "step": 1631 }, { "epoch": 1.0931011386470195, "grad_norm": 2.4947491245439473, "learning_rate": 4.5570125244955726e-06, "loss": 1.6985, "step": 1632 }, { "epoch": 1.0937709310113866, "grad_norm": 2.403642497467328, "learning_rate": 4.556181580477138e-06, "loss": 1.5687, "step": 1633 }, { "epoch": 1.0944407233757536, "grad_norm": 2.30375724830245, "learning_rate": 4.555349933765306e-06, "loss": 1.5005, "step": 1634 }, { "epoch": 1.0951105157401206, "grad_norm": 2.3020128507615514, "learning_rate": 4.554517584644288e-06, "loss": 1.5152, "step": 1635 }, { "epoch": 1.0957803081044877, "grad_norm": 3.930950630738754, "learning_rate": 4.553684533398538e-06, "loss": 1.4118, "step": 1636 }, { "epoch": 1.0964501004688547, "grad_norm": 4.3035307419272, "learning_rate": 4.552850780312747e-06, "loss": 1.3959, "step": 1637 }, { "epoch": 1.0971198928332218, "grad_norm": 4.414687630835493, "learning_rate": 4.552016325671848e-06, "loss": 1.4056, "step": 1638 }, { "epoch": 1.0977896851975888, "grad_norm": 2.448771819152584, "learning_rate": 4.5511811697610146e-06, "loss": 1.498, "step": 1639 }, { "epoch": 1.0984594775619558, "grad_norm": 2.877894760211994, "learning_rate": 4.550345312865657e-06, "loss": 1.4365, "step": 1640 }, { "epoch": 1.0991292699263229, "grad_norm": 2.1897338571617952, "learning_rate": 4.5495087552714265e-06, "loss": 1.4001, "step": 1641 }, { "epoch": 1.09979906229069, "grad_norm": 2.917563928788347, "learning_rate": 4.548671497264215e-06, "loss": 1.5067, "step": 1642 }, { "epoch": 1.100468854655057, "grad_norm": 3.645649572070861, "learning_rate": 4.547833539130152e-06, "loss": 1.5196, "step": 1643 }, { "epoch": 1.101138647019424, "grad_norm": 2.4409121272156455, "learning_rate": 4.546994881155608e-06, "loss": 1.4031, "step": 1644 }, { "epoch": 1.101808439383791, "grad_norm": 3.8745686723242474, "learning_rate": 4.54615552362719e-06, "loss": 1.5017, "step": 1645 }, { "epoch": 1.102478231748158, "grad_norm": 2.501900669525891, "learning_rate": 4.545315466831747e-06, "loss": 1.5122, "step": 1646 }, { "epoch": 1.103148024112525, "grad_norm": 2.511708145016794, "learning_rate": 4.544474711056365e-06, "loss": 1.3543, "step": 1647 }, { "epoch": 1.1038178164768921, "grad_norm": 2.6223106075161136, "learning_rate": 4.543633256588369e-06, "loss": 1.4219, "step": 1648 }, { "epoch": 1.1044876088412592, "grad_norm": 5.574557138694335, "learning_rate": 4.542791103715325e-06, "loss": 1.5379, "step": 1649 }, { "epoch": 1.1051574012056262, "grad_norm": 2.450550487209474, "learning_rate": 4.5419482527250346e-06, "loss": 1.5682, "step": 1650 }, { "epoch": 1.1058271935699933, "grad_norm": 2.4598851825528767, "learning_rate": 4.541104703905541e-06, "loss": 1.5587, "step": 1651 }, { "epoch": 1.1064969859343603, "grad_norm": 2.715963932234066, "learning_rate": 4.540260457545123e-06, "loss": 1.5164, "step": 1652 }, { "epoch": 1.1071667782987273, "grad_norm": 3.2180632155884767, "learning_rate": 4.539415513932298e-06, "loss": 1.5114, "step": 1653 }, { "epoch": 1.1078365706630944, "grad_norm": 3.012618000121958, "learning_rate": 4.5385698733558245e-06, "loss": 1.4743, "step": 1654 }, { "epoch": 1.1085063630274614, "grad_norm": 4.970887198562086, "learning_rate": 4.537723536104698e-06, "loss": 1.463, "step": 1655 }, { "epoch": 1.1091761553918285, "grad_norm": 3.013154079390049, "learning_rate": 4.536876502468149e-06, "loss": 1.3682, "step": 1656 }, { "epoch": 1.1098459477561955, "grad_norm": 5.719723834286721, "learning_rate": 4.5360287727356515e-06, "loss": 1.3941, "step": 1657 }, { "epoch": 1.1105157401205625, "grad_norm": 3.619707173833695, "learning_rate": 4.535180347196912e-06, "loss": 1.3913, "step": 1658 }, { "epoch": 1.1111855324849296, "grad_norm": 2.447639223776622, "learning_rate": 4.534331226141878e-06, "loss": 1.4984, "step": 1659 }, { "epoch": 1.1118553248492966, "grad_norm": 2.3702765514409303, "learning_rate": 4.533481409860734e-06, "loss": 1.4503, "step": 1660 }, { "epoch": 1.1125251172136639, "grad_norm": 2.9377896044303045, "learning_rate": 4.532630898643902e-06, "loss": 1.3778, "step": 1661 }, { "epoch": 1.113194909578031, "grad_norm": 2.3986053687286653, "learning_rate": 4.531779692782041e-06, "loss": 1.491, "step": 1662 }, { "epoch": 1.113864701942398, "grad_norm": 3.5266069956455826, "learning_rate": 4.5309277925660485e-06, "loss": 1.4192, "step": 1663 }, { "epoch": 1.114534494306765, "grad_norm": 2.5859071439469656, "learning_rate": 4.530075198287059e-06, "loss": 1.5211, "step": 1664 }, { "epoch": 1.115204286671132, "grad_norm": 2.362151509756807, "learning_rate": 4.529221910236442e-06, "loss": 1.7006, "step": 1665 }, { "epoch": 1.115874079035499, "grad_norm": 2.035470783865138, "learning_rate": 4.528367928705808e-06, "loss": 1.3931, "step": 1666 }, { "epoch": 1.116543871399866, "grad_norm": 2.298811806102251, "learning_rate": 4.527513253987002e-06, "loss": 1.5337, "step": 1667 }, { "epoch": 1.1172136637642331, "grad_norm": 3.0973277208977605, "learning_rate": 4.526657886372104e-06, "loss": 1.2284, "step": 1668 }, { "epoch": 1.1178834561286002, "grad_norm": 2.1894767712936445, "learning_rate": 4.525801826153436e-06, "loss": 1.4571, "step": 1669 }, { "epoch": 1.1185532484929672, "grad_norm": 3.1613594553912576, "learning_rate": 4.524945073623553e-06, "loss": 1.3755, "step": 1670 }, { "epoch": 1.1192230408573343, "grad_norm": 2.6999572978060806, "learning_rate": 4.5240876290752465e-06, "loss": 1.2783, "step": 1671 }, { "epoch": 1.1198928332217013, "grad_norm": 2.2846651590532363, "learning_rate": 4.5232294928015464e-06, "loss": 1.525, "step": 1672 }, { "epoch": 1.1205626255860683, "grad_norm": 2.601999354760743, "learning_rate": 4.522370665095718e-06, "loss": 1.6605, "step": 1673 }, { "epoch": 1.1212324179504354, "grad_norm": 2.172749027631814, "learning_rate": 4.5215111462512615e-06, "loss": 1.4932, "step": 1674 }, { "epoch": 1.1219022103148024, "grad_norm": 2.279806734083594, "learning_rate": 4.520650936561917e-06, "loss": 1.4476, "step": 1675 }, { "epoch": 1.1225720026791695, "grad_norm": 2.5149186018929566, "learning_rate": 4.519790036321656e-06, "loss": 1.2251, "step": 1676 }, { "epoch": 1.1232417950435365, "grad_norm": 2.3425411677699564, "learning_rate": 4.51892844582469e-06, "loss": 1.5882, "step": 1677 }, { "epoch": 1.1239115874079035, "grad_norm": 2.136921547359096, "learning_rate": 4.518066165365464e-06, "loss": 1.3426, "step": 1678 }, { "epoch": 1.1245813797722706, "grad_norm": 4.582094408188139, "learning_rate": 4.5172031952386596e-06, "loss": 1.3931, "step": 1679 }, { "epoch": 1.1252511721366376, "grad_norm": 2.246773522282633, "learning_rate": 4.516339535739195e-06, "loss": 1.456, "step": 1680 }, { "epoch": 1.1259209645010047, "grad_norm": 2.208929130067491, "learning_rate": 4.515475187162221e-06, "loss": 1.3873, "step": 1681 }, { "epoch": 1.1265907568653717, "grad_norm": 2.318515160721943, "learning_rate": 4.514610149803129e-06, "loss": 1.5386, "step": 1682 }, { "epoch": 1.1272605492297387, "grad_norm": 2.1042561628034706, "learning_rate": 4.51374442395754e-06, "loss": 1.2429, "step": 1683 }, { "epoch": 1.1279303415941058, "grad_norm": 3.3214882788992957, "learning_rate": 4.512878009921314e-06, "loss": 1.2738, "step": 1684 }, { "epoch": 1.1286001339584728, "grad_norm": 2.478107550726493, "learning_rate": 4.512010907990545e-06, "loss": 1.4489, "step": 1685 }, { "epoch": 1.1292699263228398, "grad_norm": 2.5203780717382718, "learning_rate": 4.511143118461562e-06, "loss": 1.2832, "step": 1686 }, { "epoch": 1.1299397186872069, "grad_norm": 2.3638415071898478, "learning_rate": 4.51027464163093e-06, "loss": 1.5296, "step": 1687 }, { "epoch": 1.130609511051574, "grad_norm": 2.172537416682814, "learning_rate": 4.509405477795448e-06, "loss": 1.5219, "step": 1688 }, { "epoch": 1.131279303415941, "grad_norm": 2.3832489228048397, "learning_rate": 4.50853562725215e-06, "loss": 1.5153, "step": 1689 }, { "epoch": 1.1319490957803082, "grad_norm": 2.464412762124149, "learning_rate": 4.507665090298303e-06, "loss": 1.4151, "step": 1690 }, { "epoch": 1.132618888144675, "grad_norm": 2.628500574004914, "learning_rate": 4.5067938672314124e-06, "loss": 1.5776, "step": 1691 }, { "epoch": 1.1332886805090423, "grad_norm": 2.4338949806309476, "learning_rate": 4.5059219583492145e-06, "loss": 1.4763, "step": 1692 }, { "epoch": 1.1339584728734093, "grad_norm": 5.397713733644475, "learning_rate": 4.505049363949683e-06, "loss": 1.5325, "step": 1693 }, { "epoch": 1.1346282652377764, "grad_norm": 2.314295189655183, "learning_rate": 4.504176084331021e-06, "loss": 1.531, "step": 1694 }, { "epoch": 1.1352980576021434, "grad_norm": 2.349821318784523, "learning_rate": 4.503302119791673e-06, "loss": 1.5328, "step": 1695 }, { "epoch": 1.1359678499665105, "grad_norm": 3.065836754536716, "learning_rate": 4.5024274706303105e-06, "loss": 1.4043, "step": 1696 }, { "epoch": 1.1366376423308775, "grad_norm": 2.7879412259465486, "learning_rate": 4.5015521371458436e-06, "loss": 1.4265, "step": 1697 }, { "epoch": 1.1373074346952445, "grad_norm": 3.7381222378160675, "learning_rate": 4.500676119637414e-06, "loss": 1.1379, "step": 1698 }, { "epoch": 1.1379772270596116, "grad_norm": 3.3311721884700334, "learning_rate": 4.499799418404398e-06, "loss": 1.365, "step": 1699 }, { "epoch": 1.1386470194239786, "grad_norm": 3.467501881196923, "learning_rate": 4.4989220337464055e-06, "loss": 1.4734, "step": 1700 }, { "epoch": 1.1393168117883457, "grad_norm": 3.1805866612516103, "learning_rate": 4.49804396596328e-06, "loss": 1.3655, "step": 1701 }, { "epoch": 1.1399866041527127, "grad_norm": 2.467842153594513, "learning_rate": 4.497165215355097e-06, "loss": 1.5884, "step": 1702 }, { "epoch": 1.1406563965170797, "grad_norm": 4.7382566468561835, "learning_rate": 4.496285782222169e-06, "loss": 1.3902, "step": 1703 }, { "epoch": 1.1413261888814468, "grad_norm": 3.148734328142864, "learning_rate": 4.495405666865038e-06, "loss": 1.5222, "step": 1704 }, { "epoch": 1.1419959812458138, "grad_norm": 2.562166220978037, "learning_rate": 4.49452486958448e-06, "loss": 1.4158, "step": 1705 }, { "epoch": 1.1426657736101808, "grad_norm": 2.0731272699969026, "learning_rate": 4.4936433906815055e-06, "loss": 1.3124, "step": 1706 }, { "epoch": 1.1433355659745479, "grad_norm": 5.289344353884968, "learning_rate": 4.492761230457358e-06, "loss": 1.3934, "step": 1707 }, { "epoch": 1.144005358338915, "grad_norm": 2.5811786629628024, "learning_rate": 4.491878389213511e-06, "loss": 1.5687, "step": 1708 }, { "epoch": 1.144675150703282, "grad_norm": 2.837411647793065, "learning_rate": 4.490994867251674e-06, "loss": 1.2975, "step": 1709 }, { "epoch": 1.145344943067649, "grad_norm": 4.0389317369558775, "learning_rate": 4.490110664873787e-06, "loss": 1.4969, "step": 1710 }, { "epoch": 1.146014735432016, "grad_norm": 3.9549073497836154, "learning_rate": 4.489225782382023e-06, "loss": 1.3015, "step": 1711 }, { "epoch": 1.146684527796383, "grad_norm": 2.390889274151135, "learning_rate": 4.488340220078789e-06, "loss": 1.4398, "step": 1712 }, { "epoch": 1.1473543201607501, "grad_norm": 2.634543677965002, "learning_rate": 4.4874539782667226e-06, "loss": 1.496, "step": 1713 }, { "epoch": 1.1480241125251172, "grad_norm": 2.472606992021558, "learning_rate": 4.486567057248693e-06, "loss": 1.6553, "step": 1714 }, { "epoch": 1.1486939048894842, "grad_norm": 4.4198381671836175, "learning_rate": 4.485679457327804e-06, "loss": 1.373, "step": 1715 }, { "epoch": 1.1493636972538512, "grad_norm": 2.5500708072723888, "learning_rate": 4.484791178807389e-06, "loss": 1.558, "step": 1716 }, { "epoch": 1.1500334896182183, "grad_norm": 2.6474447185215655, "learning_rate": 4.483902221991015e-06, "loss": 1.3663, "step": 1717 }, { "epoch": 1.1507032819825853, "grad_norm": 5.156209713988857, "learning_rate": 4.48301258718248e-06, "loss": 1.3271, "step": 1718 }, { "epoch": 1.1513730743469524, "grad_norm": 2.191874733208204, "learning_rate": 4.482122274685813e-06, "loss": 1.5425, "step": 1719 }, { "epoch": 1.1520428667113194, "grad_norm": 2.226316561105304, "learning_rate": 4.481231284805277e-06, "loss": 1.4964, "step": 1720 }, { "epoch": 1.1527126590756867, "grad_norm": 2.342662822466173, "learning_rate": 4.480339617845363e-06, "loss": 1.4904, "step": 1721 }, { "epoch": 1.1533824514400535, "grad_norm": 2.5996812611118263, "learning_rate": 4.479447274110796e-06, "loss": 1.5971, "step": 1722 }, { "epoch": 1.1540522438044207, "grad_norm": 2.188150153164989, "learning_rate": 4.478554253906533e-06, "loss": 1.4972, "step": 1723 }, { "epoch": 1.1547220361687878, "grad_norm": 3.7695352122249117, "learning_rate": 4.477660557537759e-06, "loss": 1.5015, "step": 1724 }, { "epoch": 1.1553918285331548, "grad_norm": 2.343946960666866, "learning_rate": 4.476766185309892e-06, "loss": 1.3181, "step": 1725 }, { "epoch": 1.1560616208975218, "grad_norm": 2.4430128790257344, "learning_rate": 4.475871137528582e-06, "loss": 1.4981, "step": 1726 }, { "epoch": 1.1567314132618889, "grad_norm": 2.472522995800762, "learning_rate": 4.474975414499707e-06, "loss": 1.4604, "step": 1727 }, { "epoch": 1.157401205626256, "grad_norm": 3.0980186037532333, "learning_rate": 4.47407901652938e-06, "loss": 1.4254, "step": 1728 }, { "epoch": 1.158070997990623, "grad_norm": 5.48539903187473, "learning_rate": 4.4731819439239385e-06, "loss": 1.3887, "step": 1729 }, { "epoch": 1.15874079035499, "grad_norm": 2.343462780324184, "learning_rate": 4.472284196989958e-06, "loss": 1.4716, "step": 1730 }, { "epoch": 1.159410582719357, "grad_norm": 2.2350666990729096, "learning_rate": 4.471385776034238e-06, "loss": 1.3301, "step": 1731 }, { "epoch": 1.160080375083724, "grad_norm": 2.1266388248839707, "learning_rate": 4.470486681363812e-06, "loss": 1.4488, "step": 1732 }, { "epoch": 1.1607501674480911, "grad_norm": 2.9944146845492488, "learning_rate": 4.469586913285941e-06, "loss": 1.3466, "step": 1733 }, { "epoch": 1.1614199598124582, "grad_norm": 2.372560424196351, "learning_rate": 4.4686864721081205e-06, "loss": 1.3751, "step": 1734 }, { "epoch": 1.1620897521768252, "grad_norm": 2.9115196798058287, "learning_rate": 4.467785358138072e-06, "loss": 1.1708, "step": 1735 }, { "epoch": 1.1627595445411922, "grad_norm": 2.6519098697990335, "learning_rate": 4.466883571683748e-06, "loss": 1.3911, "step": 1736 }, { "epoch": 1.1634293369055593, "grad_norm": 3.0316981305532633, "learning_rate": 4.465981113053331e-06, "loss": 1.4206, "step": 1737 }, { "epoch": 1.1640991292699263, "grad_norm": 2.3622592986374795, "learning_rate": 4.465077982555235e-06, "loss": 1.4996, "step": 1738 }, { "epoch": 1.1647689216342934, "grad_norm": 4.58395730674104, "learning_rate": 4.4641741804981e-06, "loss": 1.378, "step": 1739 }, { "epoch": 1.1654387139986604, "grad_norm": 2.6435048723816954, "learning_rate": 4.463269707190798e-06, "loss": 1.3212, "step": 1740 }, { "epoch": 1.1661085063630274, "grad_norm": 2.6116983073269977, "learning_rate": 4.462364562942431e-06, "loss": 1.1896, "step": 1741 }, { "epoch": 1.1667782987273945, "grad_norm": 5.813542575713822, "learning_rate": 4.461458748062327e-06, "loss": 1.3979, "step": 1742 }, { "epoch": 1.1674480910917615, "grad_norm": 4.063985188362161, "learning_rate": 4.460552262860046e-06, "loss": 1.4247, "step": 1743 }, { "epoch": 1.1681178834561285, "grad_norm": 2.309591333598409, "learning_rate": 4.459645107645378e-06, "loss": 1.46, "step": 1744 }, { "epoch": 1.1687876758204956, "grad_norm": 2.474235089041117, "learning_rate": 4.458737282728338e-06, "loss": 1.4142, "step": 1745 }, { "epoch": 1.1694574681848626, "grad_norm": 6.702853002229292, "learning_rate": 4.457828788419174e-06, "loss": 1.2679, "step": 1746 }, { "epoch": 1.1701272605492297, "grad_norm": 3.4726451509327694, "learning_rate": 4.45691962502836e-06, "loss": 1.2868, "step": 1747 }, { "epoch": 1.1707970529135967, "grad_norm": 5.117784142946935, "learning_rate": 4.4560097928666e-06, "loss": 1.4468, "step": 1748 }, { "epoch": 1.1714668452779637, "grad_norm": 3.7654862633148283, "learning_rate": 4.455099292244827e-06, "loss": 1.3721, "step": 1749 }, { "epoch": 1.1721366376423308, "grad_norm": 2.337772258860497, "learning_rate": 4.454188123474199e-06, "loss": 1.5666, "step": 1750 }, { "epoch": 1.1728064300066978, "grad_norm": 3.0806460858256877, "learning_rate": 4.453276286866108e-06, "loss": 1.5311, "step": 1751 }, { "epoch": 1.173476222371065, "grad_norm": 4.04553541975303, "learning_rate": 4.45236378273217e-06, "loss": 1.4203, "step": 1752 }, { "epoch": 1.174146014735432, "grad_norm": 2.42057313418194, "learning_rate": 4.45145061138423e-06, "loss": 1.3813, "step": 1753 }, { "epoch": 1.1748158070997992, "grad_norm": 4.256736301737529, "learning_rate": 4.450536773134363e-06, "loss": 1.4522, "step": 1754 }, { "epoch": 1.1754855994641662, "grad_norm": 3.371376839842629, "learning_rate": 4.449622268294868e-06, "loss": 1.4723, "step": 1755 }, { "epoch": 1.1761553918285332, "grad_norm": 2.1870416119521763, "learning_rate": 4.448707097178274e-06, "loss": 1.5381, "step": 1756 }, { "epoch": 1.1768251841929003, "grad_norm": 2.254903024283862, "learning_rate": 4.447791260097339e-06, "loss": 1.4813, "step": 1757 }, { "epoch": 1.1774949765572673, "grad_norm": 2.395920237640272, "learning_rate": 4.446874757365049e-06, "loss": 1.359, "step": 1758 }, { "epoch": 1.1781647689216344, "grad_norm": 2.519688620035255, "learning_rate": 4.445957589294611e-06, "loss": 1.4233, "step": 1759 }, { "epoch": 1.1788345612860014, "grad_norm": 2.5543296642514455, "learning_rate": 4.44503975619947e-06, "loss": 1.2787, "step": 1760 }, { "epoch": 1.1795043536503684, "grad_norm": 3.279786849230687, "learning_rate": 4.444121258393289e-06, "loss": 1.4647, "step": 1761 }, { "epoch": 1.1801741460147355, "grad_norm": 2.337129707010462, "learning_rate": 4.443202096189962e-06, "loss": 1.3584, "step": 1762 }, { "epoch": 1.1808439383791025, "grad_norm": 2.184203130918601, "learning_rate": 4.442282269903609e-06, "loss": 1.2656, "step": 1763 }, { "epoch": 1.1815137307434695, "grad_norm": 2.318972970855233, "learning_rate": 4.4413617798485795e-06, "loss": 1.344, "step": 1764 }, { "epoch": 1.1821835231078366, "grad_norm": 3.637902868811673, "learning_rate": 4.440440626339447e-06, "loss": 1.2935, "step": 1765 }, { "epoch": 1.1828533154722036, "grad_norm": 2.5102939410067187, "learning_rate": 4.439518809691013e-06, "loss": 1.4662, "step": 1766 }, { "epoch": 1.1835231078365707, "grad_norm": 2.958607698619382, "learning_rate": 4.438596330218307e-06, "loss": 1.3065, "step": 1767 }, { "epoch": 1.1841929002009377, "grad_norm": 2.252730406178419, "learning_rate": 4.43767318823658e-06, "loss": 1.5352, "step": 1768 }, { "epoch": 1.1848626925653047, "grad_norm": 7.103246018517124, "learning_rate": 4.436749384061314e-06, "loss": 1.2209, "step": 1769 }, { "epoch": 1.1855324849296718, "grad_norm": 3.4500007743806553, "learning_rate": 4.435824918008217e-06, "loss": 1.4394, "step": 1770 }, { "epoch": 1.1862022772940388, "grad_norm": 4.556869291325697, "learning_rate": 4.434899790393222e-06, "loss": 1.3652, "step": 1771 }, { "epoch": 1.1868720696584059, "grad_norm": 2.4743974479225277, "learning_rate": 4.433974001532488e-06, "loss": 1.5155, "step": 1772 }, { "epoch": 1.187541862022773, "grad_norm": 3.8595078488436703, "learning_rate": 4.433047551742401e-06, "loss": 1.3087, "step": 1773 }, { "epoch": 1.18821165438714, "grad_norm": 2.2696285639957856, "learning_rate": 4.4321204413395715e-06, "loss": 1.4301, "step": 1774 }, { "epoch": 1.188881446751507, "grad_norm": 2.692343977585916, "learning_rate": 4.431192670640836e-06, "loss": 1.59, "step": 1775 }, { "epoch": 1.189551239115874, "grad_norm": 2.3397110304634254, "learning_rate": 4.430264239963259e-06, "loss": 1.5627, "step": 1776 }, { "epoch": 1.190221031480241, "grad_norm": 2.210649824643409, "learning_rate": 4.4293351496241265e-06, "loss": 1.4438, "step": 1777 }, { "epoch": 1.190890823844608, "grad_norm": 5.069978440407623, "learning_rate": 4.428405399940954e-06, "loss": 1.4299, "step": 1778 }, { "epoch": 1.1915606162089751, "grad_norm": 2.6207985303483956, "learning_rate": 4.427474991231479e-06, "loss": 1.5164, "step": 1779 }, { "epoch": 1.1922304085733422, "grad_norm": 2.209976259894934, "learning_rate": 4.426543923813666e-06, "loss": 1.4287, "step": 1780 }, { "epoch": 1.1929002009377094, "grad_norm": 2.6018291132778764, "learning_rate": 4.425612198005704e-06, "loss": 1.6348, "step": 1781 }, { "epoch": 1.1935699933020762, "grad_norm": 2.404437856039943, "learning_rate": 4.42467981412601e-06, "loss": 1.54, "step": 1782 }, { "epoch": 1.1942397856664435, "grad_norm": 2.4281236421326553, "learning_rate": 4.423746772493219e-06, "loss": 1.5959, "step": 1783 }, { "epoch": 1.1949095780308103, "grad_norm": 2.4102570693719243, "learning_rate": 4.422813073426198e-06, "loss": 1.4668, "step": 1784 }, { "epoch": 1.1955793703951776, "grad_norm": 2.458067349889729, "learning_rate": 4.421878717244033e-06, "loss": 1.388, "step": 1785 }, { "epoch": 1.1962491627595446, "grad_norm": 2.435680657107267, "learning_rate": 4.42094370426604e-06, "loss": 1.4444, "step": 1786 }, { "epoch": 1.1969189551239117, "grad_norm": 4.027059704721161, "learning_rate": 4.420008034811756e-06, "loss": 1.3831, "step": 1787 }, { "epoch": 1.1975887474882787, "grad_norm": 4.708222491780695, "learning_rate": 4.419071709200942e-06, "loss": 1.4134, "step": 1788 }, { "epoch": 1.1982585398526457, "grad_norm": 3.3429784840084444, "learning_rate": 4.418134727753584e-06, "loss": 1.3421, "step": 1789 }, { "epoch": 1.1989283322170128, "grad_norm": 3.0992418725010484, "learning_rate": 4.417197090789893e-06, "loss": 1.4345, "step": 1790 }, { "epoch": 1.1995981245813798, "grad_norm": 2.38693074243214, "learning_rate": 4.416258798630304e-06, "loss": 1.3966, "step": 1791 }, { "epoch": 1.2002679169457469, "grad_norm": 2.483848798566516, "learning_rate": 4.415319851595474e-06, "loss": 1.4264, "step": 1792 }, { "epoch": 1.200937709310114, "grad_norm": 3.3715781076975926, "learning_rate": 4.414380250006287e-06, "loss": 1.3677, "step": 1793 }, { "epoch": 1.201607501674481, "grad_norm": 3.0037590547505633, "learning_rate": 4.413439994183847e-06, "loss": 1.4334, "step": 1794 }, { "epoch": 1.202277294038848, "grad_norm": 2.2640647912558216, "learning_rate": 4.412499084449484e-06, "loss": 1.4845, "step": 1795 }, { "epoch": 1.202947086403215, "grad_norm": 3.1829090859053197, "learning_rate": 4.411557521124751e-06, "loss": 1.4516, "step": 1796 }, { "epoch": 1.203616878767582, "grad_norm": 2.1268086450555566, "learning_rate": 4.410615304531424e-06, "loss": 1.3908, "step": 1797 }, { "epoch": 1.204286671131949, "grad_norm": 2.4729379015528523, "learning_rate": 4.409672434991503e-06, "loss": 1.4462, "step": 1798 }, { "epoch": 1.2049564634963161, "grad_norm": 2.036128875827382, "learning_rate": 4.4087289128272095e-06, "loss": 1.3238, "step": 1799 }, { "epoch": 1.2056262558606832, "grad_norm": 2.572239153168522, "learning_rate": 4.407784738360991e-06, "loss": 1.428, "step": 1800 }, { "epoch": 1.2062960482250502, "grad_norm": 2.6134060373284735, "learning_rate": 4.406839911915513e-06, "loss": 1.3111, "step": 1801 }, { "epoch": 1.2069658405894172, "grad_norm": 2.529373606037773, "learning_rate": 4.405894433813671e-06, "loss": 1.3937, "step": 1802 }, { "epoch": 1.2076356329537843, "grad_norm": 2.295433268757157, "learning_rate": 4.404948304378575e-06, "loss": 1.5894, "step": 1803 }, { "epoch": 1.2083054253181513, "grad_norm": 2.583046265966884, "learning_rate": 4.404001523933564e-06, "loss": 1.5466, "step": 1804 }, { "epoch": 1.2089752176825184, "grad_norm": 3.2175616705284726, "learning_rate": 4.403054092802198e-06, "loss": 1.3138, "step": 1805 }, { "epoch": 1.2096450100468854, "grad_norm": 2.319124629520042, "learning_rate": 4.402106011308256e-06, "loss": 1.2497, "step": 1806 }, { "epoch": 1.2103148024112524, "grad_norm": 4.4899976620029545, "learning_rate": 4.401157279775743e-06, "loss": 1.37, "step": 1807 }, { "epoch": 1.2109845947756195, "grad_norm": 2.6917598061182204, "learning_rate": 4.400207898528885e-06, "loss": 1.311, "step": 1808 }, { "epoch": 1.2116543871399865, "grad_norm": 2.610760089005322, "learning_rate": 4.399257867892131e-06, "loss": 0.932, "step": 1809 }, { "epoch": 1.2123241795043536, "grad_norm": 5.956454526576308, "learning_rate": 4.398307188190149e-06, "loss": 1.4582, "step": 1810 }, { "epoch": 1.2129939718687206, "grad_norm": 2.6368597300468064, "learning_rate": 4.397355859747831e-06, "loss": 1.3662, "step": 1811 }, { "epoch": 1.2136637642330879, "grad_norm": 2.3232235039764686, "learning_rate": 4.39640388289029e-06, "loss": 1.4876, "step": 1812 }, { "epoch": 1.2143335565974547, "grad_norm": 2.0828576763093736, "learning_rate": 4.395451257942864e-06, "loss": 1.4292, "step": 1813 }, { "epoch": 1.215003348961822, "grad_norm": 2.2545022872256957, "learning_rate": 4.394497985231107e-06, "loss": 1.437, "step": 1814 }, { "epoch": 1.215673141326189, "grad_norm": 2.4850346019991565, "learning_rate": 4.3935440650807955e-06, "loss": 1.379, "step": 1815 }, { "epoch": 1.216342933690556, "grad_norm": 3.3186356053250514, "learning_rate": 4.392589497817933e-06, "loss": 1.5405, "step": 1816 }, { "epoch": 1.217012726054923, "grad_norm": 3.147063407720659, "learning_rate": 4.391634283768735e-06, "loss": 1.3581, "step": 1817 }, { "epoch": 1.21768251841929, "grad_norm": 2.22035550922312, "learning_rate": 4.3906784232596464e-06, "loss": 1.3076, "step": 1818 }, { "epoch": 1.2183523107836571, "grad_norm": 2.2445388531151935, "learning_rate": 4.389721916617328e-06, "loss": 1.3541, "step": 1819 }, { "epoch": 1.2190221031480242, "grad_norm": 2.150254429127932, "learning_rate": 4.388764764168661e-06, "loss": 1.3434, "step": 1820 }, { "epoch": 1.2196918955123912, "grad_norm": 2.4367261968354588, "learning_rate": 4.387806966240753e-06, "loss": 1.4484, "step": 1821 }, { "epoch": 1.2203616878767582, "grad_norm": 2.3328084627620123, "learning_rate": 4.386848523160926e-06, "loss": 1.3708, "step": 1822 }, { "epoch": 1.2210314802411253, "grad_norm": 2.3365829085663434, "learning_rate": 4.385889435256725e-06, "loss": 1.5493, "step": 1823 }, { "epoch": 1.2217012726054923, "grad_norm": 2.7934487024019923, "learning_rate": 4.384929702855915e-06, "loss": 1.5508, "step": 1824 }, { "epoch": 1.2223710649698594, "grad_norm": 3.565757657785549, "learning_rate": 4.3839693262864816e-06, "loss": 1.0486, "step": 1825 }, { "epoch": 1.2230408573342264, "grad_norm": 3.147932658712464, "learning_rate": 4.383008305876632e-06, "loss": 1.2753, "step": 1826 }, { "epoch": 1.2237106496985934, "grad_norm": 3.146057491514247, "learning_rate": 4.382046641954789e-06, "loss": 1.4487, "step": 1827 }, { "epoch": 1.2243804420629605, "grad_norm": 2.8038833181533134, "learning_rate": 4.3810843348496e-06, "loss": 1.4118, "step": 1828 }, { "epoch": 1.2250502344273275, "grad_norm": 2.3811801733667375, "learning_rate": 4.3801213848899295e-06, "loss": 1.3322, "step": 1829 }, { "epoch": 1.2257200267916946, "grad_norm": 2.233285796262412, "learning_rate": 4.379157792404864e-06, "loss": 1.3902, "step": 1830 }, { "epoch": 1.2263898191560616, "grad_norm": 2.5588732797282248, "learning_rate": 4.378193557723706e-06, "loss": 1.3982, "step": 1831 }, { "epoch": 1.2270596115204286, "grad_norm": 2.8763386701572498, "learning_rate": 4.37722868117598e-06, "loss": 1.4327, "step": 1832 }, { "epoch": 1.2277294038847957, "grad_norm": 2.253227934692992, "learning_rate": 4.376263163091431e-06, "loss": 1.4346, "step": 1833 }, { "epoch": 1.2283991962491627, "grad_norm": 2.5079405761970834, "learning_rate": 4.375297003800021e-06, "loss": 1.305, "step": 1834 }, { "epoch": 1.2290689886135298, "grad_norm": 2.699506602511993, "learning_rate": 4.37433020363193e-06, "loss": 1.3279, "step": 1835 }, { "epoch": 1.2297387809778968, "grad_norm": 2.100743856131612, "learning_rate": 4.373362762917561e-06, "loss": 1.4377, "step": 1836 }, { "epoch": 1.2304085733422638, "grad_norm": 2.3911864925099646, "learning_rate": 4.372394681987534e-06, "loss": 1.5322, "step": 1837 }, { "epoch": 1.2310783657066309, "grad_norm": 2.34728717270343, "learning_rate": 4.371425961172686e-06, "loss": 1.4131, "step": 1838 }, { "epoch": 1.231748158070998, "grad_norm": 2.3251224191230833, "learning_rate": 4.370456600804075e-06, "loss": 1.547, "step": 1839 }, { "epoch": 1.232417950435365, "grad_norm": 2.4743913242428883, "learning_rate": 4.369486601212976e-06, "loss": 1.498, "step": 1840 }, { "epoch": 1.233087742799732, "grad_norm": 4.145581009103769, "learning_rate": 4.368515962730883e-06, "loss": 1.1964, "step": 1841 }, { "epoch": 1.233757535164099, "grad_norm": 3.9461079359897533, "learning_rate": 4.367544685689511e-06, "loss": 1.3475, "step": 1842 }, { "epoch": 1.2344273275284663, "grad_norm": 2.266015934757842, "learning_rate": 4.36657277042079e-06, "loss": 1.367, "step": 1843 }, { "epoch": 1.235097119892833, "grad_norm": 2.3438131817023478, "learning_rate": 4.365600217256866e-06, "loss": 1.3672, "step": 1844 }, { "epoch": 1.2357669122572004, "grad_norm": 2.3131943709237066, "learning_rate": 4.364627026530108e-06, "loss": 1.5547, "step": 1845 }, { "epoch": 1.2364367046215674, "grad_norm": 2.8357073226021923, "learning_rate": 4.363653198573101e-06, "loss": 1.2978, "step": 1846 }, { "epoch": 1.2371064969859344, "grad_norm": 3.070557036571916, "learning_rate": 4.362678733718649e-06, "loss": 1.3765, "step": 1847 }, { "epoch": 1.2377762893503015, "grad_norm": 2.4601776072507717, "learning_rate": 4.361703632299768e-06, "loss": 1.2733, "step": 1848 }, { "epoch": 1.2384460817146685, "grad_norm": 2.8170181384658077, "learning_rate": 4.360727894649699e-06, "loss": 1.4585, "step": 1849 }, { "epoch": 1.2391158740790356, "grad_norm": 2.529907839012991, "learning_rate": 4.359751521101896e-06, "loss": 1.5074, "step": 1850 }, { "epoch": 1.2397856664434026, "grad_norm": 2.9525965080055294, "learning_rate": 4.3587745119900325e-06, "loss": 1.3435, "step": 1851 }, { "epoch": 1.2404554588077696, "grad_norm": 2.941154761380468, "learning_rate": 4.357796867647998e-06, "loss": 1.4592, "step": 1852 }, { "epoch": 1.2411252511721367, "grad_norm": 2.39575873468641, "learning_rate": 4.356818588409898e-06, "loss": 1.4309, "step": 1853 }, { "epoch": 1.2417950435365037, "grad_norm": 2.3206374974995247, "learning_rate": 4.355839674610058e-06, "loss": 1.4329, "step": 1854 }, { "epoch": 1.2424648359008708, "grad_norm": 2.358276744463772, "learning_rate": 4.354860126583017e-06, "loss": 1.4623, "step": 1855 }, { "epoch": 1.2431346282652378, "grad_norm": 4.647882755820502, "learning_rate": 4.3538799446635335e-06, "loss": 1.4909, "step": 1856 }, { "epoch": 1.2438044206296048, "grad_norm": 2.7576708623770267, "learning_rate": 4.352899129186581e-06, "loss": 1.4057, "step": 1857 }, { "epoch": 1.2444742129939719, "grad_norm": 2.701613316186023, "learning_rate": 4.351917680487351e-06, "loss": 1.4196, "step": 1858 }, { "epoch": 1.245144005358339, "grad_norm": 2.750092071570625, "learning_rate": 4.350935598901248e-06, "loss": 1.4988, "step": 1859 }, { "epoch": 1.245813797722706, "grad_norm": 2.740127104248417, "learning_rate": 4.349952884763898e-06, "loss": 1.3777, "step": 1860 }, { "epoch": 1.246483590087073, "grad_norm": 2.508521896805711, "learning_rate": 4.348969538411138e-06, "loss": 1.3829, "step": 1861 }, { "epoch": 1.24715338245144, "grad_norm": 2.513970812459005, "learning_rate": 4.347985560179025e-06, "loss": 1.402, "step": 1862 }, { "epoch": 1.247823174815807, "grad_norm": 5.152239902848764, "learning_rate": 4.347000950403831e-06, "loss": 1.4218, "step": 1863 }, { "epoch": 1.248492967180174, "grad_norm": 2.6690660326091162, "learning_rate": 4.3460157094220415e-06, "loss": 1.5776, "step": 1864 }, { "epoch": 1.2491627595445411, "grad_norm": 2.642091521655077, "learning_rate": 4.345029837570359e-06, "loss": 1.5341, "step": 1865 }, { "epoch": 1.2498325519089082, "grad_norm": 3.7562588280465943, "learning_rate": 4.344043335185705e-06, "loss": 1.4379, "step": 1866 }, { "epoch": 1.2505023442732752, "grad_norm": 3.351273981845172, "learning_rate": 4.34305620260521e-06, "loss": 1.6409, "step": 1867 }, { "epoch": 1.2511721366376423, "grad_norm": 2.475615826946806, "learning_rate": 4.342068440166225e-06, "loss": 1.3208, "step": 1868 }, { "epoch": 1.2518419290020093, "grad_norm": 4.1255985290637796, "learning_rate": 4.3410800482063155e-06, "loss": 1.3978, "step": 1869 }, { "epoch": 1.2525117213663763, "grad_norm": 2.782854643461701, "learning_rate": 4.34009102706326e-06, "loss": 1.4162, "step": 1870 }, { "epoch": 1.2531815137307434, "grad_norm": 3.562962481478877, "learning_rate": 4.339101377075054e-06, "loss": 1.406, "step": 1871 }, { "epoch": 1.2538513060951106, "grad_norm": 6.268709458958145, "learning_rate": 4.338111098579905e-06, "loss": 1.1944, "step": 1872 }, { "epoch": 1.2545210984594775, "grad_norm": 2.957211906787835, "learning_rate": 4.337120191916241e-06, "loss": 1.4155, "step": 1873 }, { "epoch": 1.2551908908238447, "grad_norm": 2.457922957536186, "learning_rate": 4.336128657422698e-06, "loss": 1.3625, "step": 1874 }, { "epoch": 1.2558606831882115, "grad_norm": 3.498713204611035, "learning_rate": 4.335136495438132e-06, "loss": 1.5392, "step": 1875 }, { "epoch": 1.2565304755525788, "grad_norm": 2.51818965953288, "learning_rate": 4.33414370630161e-06, "loss": 1.442, "step": 1876 }, { "epoch": 1.2572002679169456, "grad_norm": 2.6023789360218554, "learning_rate": 4.333150290352415e-06, "loss": 1.5445, "step": 1877 }, { "epoch": 1.2578700602813129, "grad_norm": 2.4559271715545203, "learning_rate": 4.332156247930043e-06, "loss": 1.2781, "step": 1878 }, { "epoch": 1.25853985264568, "grad_norm": 2.431126645924794, "learning_rate": 4.331161579374205e-06, "loss": 1.5383, "step": 1879 }, { "epoch": 1.259209645010047, "grad_norm": 3.306574923758327, "learning_rate": 4.330166285024827e-06, "loss": 1.4127, "step": 1880 }, { "epoch": 1.259879437374414, "grad_norm": 2.9709605239310433, "learning_rate": 4.329170365222045e-06, "loss": 1.372, "step": 1881 }, { "epoch": 1.260549229738781, "grad_norm": 3.206883615539777, "learning_rate": 4.328173820306214e-06, "loss": 1.2948, "step": 1882 }, { "epoch": 1.261219022103148, "grad_norm": 2.420595736277939, "learning_rate": 4.327176650617898e-06, "loss": 1.2806, "step": 1883 }, { "epoch": 1.261888814467515, "grad_norm": 2.250128174125947, "learning_rate": 4.326178856497878e-06, "loss": 1.3327, "step": 1884 }, { "epoch": 1.2625586068318821, "grad_norm": 7.1933995384564, "learning_rate": 4.325180438287147e-06, "loss": 1.2112, "step": 1885 }, { "epoch": 1.2632283991962492, "grad_norm": 3.2779327771906663, "learning_rate": 4.324181396326909e-06, "loss": 1.5208, "step": 1886 }, { "epoch": 1.2638981915606162, "grad_norm": 2.5093046559938976, "learning_rate": 4.3231817309585855e-06, "loss": 1.5706, "step": 1887 }, { "epoch": 1.2645679839249833, "grad_norm": 2.321851209778937, "learning_rate": 4.322181442523809e-06, "loss": 1.1236, "step": 1888 }, { "epoch": 1.2652377762893503, "grad_norm": 2.298643699958479, "learning_rate": 4.321180531364423e-06, "loss": 1.4763, "step": 1889 }, { "epoch": 1.2659075686537173, "grad_norm": 3.500422069840603, "learning_rate": 4.320178997822487e-06, "loss": 1.4809, "step": 1890 }, { "epoch": 1.2665773610180844, "grad_norm": 3.6886216056461927, "learning_rate": 4.3191768422402725e-06, "loss": 1.4646, "step": 1891 }, { "epoch": 1.2672471533824514, "grad_norm": 2.3535159649178645, "learning_rate": 4.318174064960261e-06, "loss": 1.5425, "step": 1892 }, { "epoch": 1.2679169457468185, "grad_norm": 4.401782671994098, "learning_rate": 4.31717066632515e-06, "loss": 1.5147, "step": 1893 }, { "epoch": 1.2685867381111855, "grad_norm": 2.7934003657969977, "learning_rate": 4.316166646677847e-06, "loss": 1.1547, "step": 1894 }, { "epoch": 1.2692565304755525, "grad_norm": 3.30930352647628, "learning_rate": 4.315162006361472e-06, "loss": 1.2825, "step": 1895 }, { "epoch": 1.2699263228399196, "grad_norm": 4.0290204403758025, "learning_rate": 4.314156745719359e-06, "loss": 1.4026, "step": 1896 }, { "epoch": 1.2705961152042866, "grad_norm": 4.069654968965371, "learning_rate": 4.313150865095052e-06, "loss": 1.4452, "step": 1897 }, { "epoch": 1.2712659075686537, "grad_norm": 2.5571699091045663, "learning_rate": 4.312144364832307e-06, "loss": 1.3301, "step": 1898 }, { "epoch": 1.2719356999330207, "grad_norm": 2.2921399504754723, "learning_rate": 4.311137245275091e-06, "loss": 1.3709, "step": 1899 }, { "epoch": 1.2726054922973877, "grad_norm": 2.4410804347150017, "learning_rate": 4.310129506767587e-06, "loss": 1.4715, "step": 1900 }, { "epoch": 1.273275284661755, "grad_norm": 3.221532744127578, "learning_rate": 4.309121149654184e-06, "loss": 1.2796, "step": 1901 }, { "epoch": 1.2739450770261218, "grad_norm": 2.284018963922588, "learning_rate": 4.308112174279485e-06, "loss": 1.4502, "step": 1902 }, { "epoch": 1.274614869390489, "grad_norm": 3.4183599826059874, "learning_rate": 4.307102580988305e-06, "loss": 1.4047, "step": 1903 }, { "epoch": 1.2752846617548559, "grad_norm": 2.63253682272127, "learning_rate": 4.3060923701256686e-06, "loss": 1.4947, "step": 1904 }, { "epoch": 1.2759544541192231, "grad_norm": 2.4756467563745144, "learning_rate": 4.305081542036813e-06, "loss": 1.464, "step": 1905 }, { "epoch": 1.27662424648359, "grad_norm": 2.1447891626367808, "learning_rate": 4.3040700970671835e-06, "loss": 1.3412, "step": 1906 }, { "epoch": 1.2772940388479572, "grad_norm": 2.4711978800483805, "learning_rate": 4.30305803556244e-06, "loss": 1.668, "step": 1907 }, { "epoch": 1.277963831212324, "grad_norm": 2.754823481883857, "learning_rate": 4.30204535786845e-06, "loss": 1.3825, "step": 1908 }, { "epoch": 1.2786336235766913, "grad_norm": 2.5818282205151935, "learning_rate": 4.3010320643312935e-06, "loss": 1.3523, "step": 1909 }, { "epoch": 1.2793034159410583, "grad_norm": 3.1370807518978627, "learning_rate": 4.30001815529726e-06, "loss": 1.5564, "step": 1910 }, { "epoch": 1.2799732083054254, "grad_norm": 3.4768411428102217, "learning_rate": 4.299003631112849e-06, "loss": 1.587, "step": 1911 }, { "epoch": 1.2806430006697924, "grad_norm": 2.3392370722528177, "learning_rate": 4.2979884921247726e-06, "loss": 1.3201, "step": 1912 }, { "epoch": 1.2813127930341595, "grad_norm": 11.173352876021692, "learning_rate": 4.296972738679951e-06, "loss": 1.3203, "step": 1913 }, { "epoch": 1.2819825853985265, "grad_norm": 3.106932076332443, "learning_rate": 4.295956371125513e-06, "loss": 1.3834, "step": 1914 }, { "epoch": 1.2826523777628935, "grad_norm": 2.383549594123683, "learning_rate": 4.2949393898088e-06, "loss": 1.4287, "step": 1915 }, { "epoch": 1.2833221701272606, "grad_norm": 2.286316375586928, "learning_rate": 4.293921795077361e-06, "loss": 1.6014, "step": 1916 }, { "epoch": 1.2839919624916276, "grad_norm": 2.3506755416017597, "learning_rate": 4.292903587278958e-06, "loss": 1.341, "step": 1917 }, { "epoch": 1.2846617548559947, "grad_norm": 2.3078019020751714, "learning_rate": 4.291884766761558e-06, "loss": 1.4772, "step": 1918 }, { "epoch": 1.2853315472203617, "grad_norm": 2.3397628216444546, "learning_rate": 4.290865333873341e-06, "loss": 1.282, "step": 1919 }, { "epoch": 1.2860013395847287, "grad_norm": 2.3591611044599943, "learning_rate": 4.289845288962694e-06, "loss": 1.3505, "step": 1920 }, { "epoch": 1.2866711319490958, "grad_norm": 2.5306743312870905, "learning_rate": 4.2888246323782145e-06, "loss": 1.2245, "step": 1921 }, { "epoch": 1.2873409243134628, "grad_norm": 3.512618550396618, "learning_rate": 4.287803364468709e-06, "loss": 1.3635, "step": 1922 }, { "epoch": 1.2880107166778298, "grad_norm": 3.235155713022266, "learning_rate": 4.286781485583191e-06, "loss": 1.4477, "step": 1923 }, { "epoch": 1.2886805090421969, "grad_norm": 2.795029420612409, "learning_rate": 4.285758996070886e-06, "loss": 1.1568, "step": 1924 }, { "epoch": 1.289350301406564, "grad_norm": 3.4831789993292834, "learning_rate": 4.2847358962812255e-06, "loss": 1.379, "step": 1925 }, { "epoch": 1.290020093770931, "grad_norm": 3.2379713806571524, "learning_rate": 4.283712186563851e-06, "loss": 1.287, "step": 1926 }, { "epoch": 1.290689886135298, "grad_norm": 3.1335003447196286, "learning_rate": 4.282687867268612e-06, "loss": 1.4513, "step": 1927 }, { "epoch": 1.291359678499665, "grad_norm": 3.732722463217025, "learning_rate": 4.281662938745565e-06, "loss": 1.4082, "step": 1928 }, { "epoch": 1.292029470864032, "grad_norm": 2.4549029579915653, "learning_rate": 4.280637401344978e-06, "loss": 1.4084, "step": 1929 }, { "epoch": 1.2926992632283991, "grad_norm": 2.3411723805832874, "learning_rate": 4.279611255417324e-06, "loss": 1.5483, "step": 1930 }, { "epoch": 1.2933690555927662, "grad_norm": 2.2677924548530752, "learning_rate": 4.278584501313286e-06, "loss": 1.4548, "step": 1931 }, { "epoch": 1.2940388479571334, "grad_norm": 2.1849201116759693, "learning_rate": 4.277557139383752e-06, "loss": 1.4257, "step": 1932 }, { "epoch": 1.2947086403215002, "grad_norm": 3.3400211627175853, "learning_rate": 4.2765291699798205e-06, "loss": 1.3663, "step": 1933 }, { "epoch": 1.2953784326858675, "grad_norm": 2.769178180687636, "learning_rate": 4.275500593452798e-06, "loss": 1.1878, "step": 1934 }, { "epoch": 1.2960482250502343, "grad_norm": 3.1923534405563996, "learning_rate": 4.274471410154195e-06, "loss": 1.4646, "step": 1935 }, { "epoch": 1.2967180174146016, "grad_norm": 3.5172825501529195, "learning_rate": 4.2734416204357335e-06, "loss": 1.2373, "step": 1936 }, { "epoch": 1.2973878097789684, "grad_norm": 2.481097731377914, "learning_rate": 4.27241122464934e-06, "loss": 1.345, "step": 1937 }, { "epoch": 1.2980576021433357, "grad_norm": 3.6170288417046765, "learning_rate": 4.271380223147149e-06, "loss": 1.2616, "step": 1938 }, { "epoch": 1.2987273945077027, "grad_norm": 2.840105227331563, "learning_rate": 4.270348616281501e-06, "loss": 1.4305, "step": 1939 }, { "epoch": 1.2993971868720697, "grad_norm": 2.5651136221499726, "learning_rate": 4.269316404404946e-06, "loss": 1.3118, "step": 1940 }, { "epoch": 1.3000669792364368, "grad_norm": 2.3711547409907587, "learning_rate": 4.2682835878702376e-06, "loss": 1.4794, "step": 1941 }, { "epoch": 1.3007367716008038, "grad_norm": 4.065630216715508, "learning_rate": 4.267250167030339e-06, "loss": 1.3261, "step": 1942 }, { "epoch": 1.3014065639651708, "grad_norm": 2.5909524969971307, "learning_rate": 4.266216142238417e-06, "loss": 1.166, "step": 1943 }, { "epoch": 1.3020763563295379, "grad_norm": 2.281092448305185, "learning_rate": 4.2651815138478465e-06, "loss": 1.3696, "step": 1944 }, { "epoch": 1.302746148693905, "grad_norm": 2.390686103531768, "learning_rate": 4.264146282212209e-06, "loss": 1.4566, "step": 1945 }, { "epoch": 1.303415941058272, "grad_norm": 2.634184512565769, "learning_rate": 4.263110447685291e-06, "loss": 0.9142, "step": 1946 }, { "epoch": 1.304085733422639, "grad_norm": 3.348134774101887, "learning_rate": 4.262074010621087e-06, "loss": 1.4627, "step": 1947 }, { "epoch": 1.304755525787006, "grad_norm": 3.3293057528858774, "learning_rate": 4.261036971373794e-06, "loss": 1.2556, "step": 1948 }, { "epoch": 1.305425318151373, "grad_norm": 2.7753662757945956, "learning_rate": 4.259999330297817e-06, "loss": 1.4921, "step": 1949 }, { "epoch": 1.3060951105157401, "grad_norm": 2.127363205286999, "learning_rate": 4.2589610877477685e-06, "loss": 1.4843, "step": 1950 }, { "epoch": 1.3067649028801072, "grad_norm": 2.202335387071879, "learning_rate": 4.2579222440784625e-06, "loss": 1.3958, "step": 1951 }, { "epoch": 1.3074346952444742, "grad_norm": 2.180790921325842, "learning_rate": 4.2568827996449205e-06, "loss": 1.159, "step": 1952 }, { "epoch": 1.3081044876088412, "grad_norm": 3.361186775970189, "learning_rate": 4.2558427548023704e-06, "loss": 1.4862, "step": 1953 }, { "epoch": 1.3087742799732083, "grad_norm": 3.1403263651222466, "learning_rate": 4.254802109906244e-06, "loss": 1.3789, "step": 1954 }, { "epoch": 1.3094440723375753, "grad_norm": 2.3009543575221296, "learning_rate": 4.253760865312178e-06, "loss": 1.2684, "step": 1955 }, { "epoch": 1.3101138647019424, "grad_norm": 2.8149853333522787, "learning_rate": 4.252719021376014e-06, "loss": 1.3335, "step": 1956 }, { "epoch": 1.3107836570663094, "grad_norm": 6.935143864259581, "learning_rate": 4.2516765784537994e-06, "loss": 1.3136, "step": 1957 }, { "epoch": 1.3114534494306764, "grad_norm": 2.4611996209634417, "learning_rate": 4.2506335369017844e-06, "loss": 1.3098, "step": 1958 }, { "epoch": 1.3121232417950435, "grad_norm": 3.8611084535502003, "learning_rate": 4.249589897076427e-06, "loss": 1.5712, "step": 1959 }, { "epoch": 1.3127930341594105, "grad_norm": 3.315334899658991, "learning_rate": 4.248545659334386e-06, "loss": 1.3976, "step": 1960 }, { "epoch": 1.3134628265237775, "grad_norm": 4.548387981680201, "learning_rate": 4.247500824032529e-06, "loss": 1.2262, "step": 1961 }, { "epoch": 1.3141326188881446, "grad_norm": 2.5722865355592184, "learning_rate": 4.246455391527921e-06, "loss": 1.2534, "step": 1962 }, { "epoch": 1.3148024112525118, "grad_norm": 2.266954775859108, "learning_rate": 4.2454093621778396e-06, "loss": 1.5812, "step": 1963 }, { "epoch": 1.3154722036168787, "grad_norm": 2.4251790627405523, "learning_rate": 4.244362736339758e-06, "loss": 1.3634, "step": 1964 }, { "epoch": 1.316141995981246, "grad_norm": 2.5533574410527264, "learning_rate": 4.243315514371359e-06, "loss": 1.3449, "step": 1965 }, { "epoch": 1.3168117883456127, "grad_norm": 3.5751873490529382, "learning_rate": 4.2422676966305285e-06, "loss": 1.6046, "step": 1966 }, { "epoch": 1.31748158070998, "grad_norm": 3.7674914520813587, "learning_rate": 4.241219283475352e-06, "loss": 1.4111, "step": 1967 }, { "epoch": 1.3181513730743468, "grad_norm": 2.521777948041008, "learning_rate": 4.240170275264124e-06, "loss": 1.623, "step": 1968 }, { "epoch": 1.318821165438714, "grad_norm": 2.1288782365461545, "learning_rate": 4.239120672355338e-06, "loss": 1.346, "step": 1969 }, { "epoch": 1.3194909578030811, "grad_norm": 3.7259234646869577, "learning_rate": 4.2380704751076925e-06, "loss": 1.2057, "step": 1970 }, { "epoch": 1.3201607501674482, "grad_norm": 2.2699192252919036, "learning_rate": 4.23701968388009e-06, "loss": 1.377, "step": 1971 }, { "epoch": 1.3208305425318152, "grad_norm": 2.222252683193258, "learning_rate": 4.235968299031634e-06, "loss": 1.3704, "step": 1972 }, { "epoch": 1.3215003348961822, "grad_norm": 3.1113402240279293, "learning_rate": 4.234916320921633e-06, "loss": 1.4337, "step": 1973 }, { "epoch": 1.3221701272605493, "grad_norm": 3.3215164068123335, "learning_rate": 4.2338637499095956e-06, "loss": 1.3357, "step": 1974 }, { "epoch": 1.3228399196249163, "grad_norm": 2.885014108328046, "learning_rate": 4.232810586355235e-06, "loss": 1.4018, "step": 1975 }, { "epoch": 1.3235097119892834, "grad_norm": 2.3639292842971535, "learning_rate": 4.231756830618467e-06, "loss": 1.4675, "step": 1976 }, { "epoch": 1.3241795043536504, "grad_norm": 2.4884790579200278, "learning_rate": 4.23070248305941e-06, "loss": 1.2937, "step": 1977 }, { "epoch": 1.3248492967180174, "grad_norm": 2.468919610313452, "learning_rate": 4.229647544038382e-06, "loss": 1.4273, "step": 1978 }, { "epoch": 1.3255190890823845, "grad_norm": 2.4194699865652094, "learning_rate": 4.228592013915905e-06, "loss": 1.4855, "step": 1979 }, { "epoch": 1.3261888814467515, "grad_norm": 2.5191479177399754, "learning_rate": 4.227535893052705e-06, "loss": 1.5001, "step": 1980 }, { "epoch": 1.3268586738111185, "grad_norm": 2.317818459051653, "learning_rate": 4.2264791818097075e-06, "loss": 1.4773, "step": 1981 }, { "epoch": 1.3275284661754856, "grad_norm": 2.3895610395067766, "learning_rate": 4.225421880548038e-06, "loss": 1.2273, "step": 1982 }, { "epoch": 1.3281982585398526, "grad_norm": 3.4235129146753107, "learning_rate": 4.2243639896290264e-06, "loss": 1.2839, "step": 1983 }, { "epoch": 1.3288680509042197, "grad_norm": 5.68636537319073, "learning_rate": 4.223305509414206e-06, "loss": 1.3241, "step": 1984 }, { "epoch": 1.3295378432685867, "grad_norm": 2.1196963932441695, "learning_rate": 4.222246440265306e-06, "loss": 1.2059, "step": 1985 }, { "epoch": 1.3302076356329537, "grad_norm": 3.05872173511463, "learning_rate": 4.221186782544262e-06, "loss": 1.5259, "step": 1986 }, { "epoch": 1.3308774279973208, "grad_norm": 2.5513057218085304, "learning_rate": 4.220126536613207e-06, "loss": 1.4231, "step": 1987 }, { "epoch": 1.3315472203616878, "grad_norm": 2.2088917870076883, "learning_rate": 4.219065702834478e-06, "loss": 1.3336, "step": 1988 }, { "epoch": 1.3322170127260549, "grad_norm": 2.0419343576281794, "learning_rate": 4.21800428157061e-06, "loss": 1.2521, "step": 1989 }, { "epoch": 1.332886805090422, "grad_norm": 3.325849707017833, "learning_rate": 4.216942273184341e-06, "loss": 1.2338, "step": 1990 }, { "epoch": 1.333556597454789, "grad_norm": 2.301403695337683, "learning_rate": 4.215879678038609e-06, "loss": 1.4318, "step": 1991 }, { "epoch": 1.3342263898191562, "grad_norm": 3.392272963369058, "learning_rate": 4.214816496496553e-06, "loss": 1.4574, "step": 1992 }, { "epoch": 1.334896182183523, "grad_norm": 2.5611090208709477, "learning_rate": 4.2137527289215115e-06, "loss": 1.3172, "step": 1993 }, { "epoch": 1.3355659745478903, "grad_norm": 3.882838140772493, "learning_rate": 4.2126883756770235e-06, "loss": 1.3394, "step": 1994 }, { "epoch": 1.336235766912257, "grad_norm": 4.030673418955701, "learning_rate": 4.211623437126827e-06, "loss": 1.3564, "step": 1995 }, { "epoch": 1.3369055592766244, "grad_norm": 2.261568645541099, "learning_rate": 4.210557913634864e-06, "loss": 1.1849, "step": 1996 }, { "epoch": 1.3375753516409912, "grad_norm": 2.463796204508042, "learning_rate": 4.209491805565271e-06, "loss": 1.2479, "step": 1997 }, { "epoch": 1.3382451440053584, "grad_norm": 2.230691961096254, "learning_rate": 4.208425113282388e-06, "loss": 1.3423, "step": 1998 }, { "epoch": 1.3389149363697253, "grad_norm": 2.7173342145595285, "learning_rate": 4.207357837150755e-06, "loss": 1.2713, "step": 1999 }, { "epoch": 1.3395847287340925, "grad_norm": 3.2600040930775274, "learning_rate": 4.206289977535108e-06, "loss": 1.4257, "step": 2000 }, { "epoch": 1.3402545210984596, "grad_norm": 3.699518517378992, "learning_rate": 4.205221534800386e-06, "loss": 1.4694, "step": 2001 }, { "epoch": 1.3409243134628266, "grad_norm": 2.9805174065676634, "learning_rate": 4.204152509311726e-06, "loss": 1.21, "step": 2002 }, { "epoch": 1.3415941058271936, "grad_norm": 2.2510568148052017, "learning_rate": 4.203082901434463e-06, "loss": 1.4944, "step": 2003 }, { "epoch": 1.3422638981915607, "grad_norm": 3.279259928551706, "learning_rate": 4.202012711534132e-06, "loss": 1.3139, "step": 2004 }, { "epoch": 1.3429336905559277, "grad_norm": 2.9689558546222146, "learning_rate": 4.200941939976467e-06, "loss": 1.4133, "step": 2005 }, { "epoch": 1.3436034829202947, "grad_norm": 2.2806331321714293, "learning_rate": 4.199870587127402e-06, "loss": 1.2104, "step": 2006 }, { "epoch": 1.3442732752846618, "grad_norm": 2.709778729583334, "learning_rate": 4.1987986533530675e-06, "loss": 1.3528, "step": 2007 }, { "epoch": 1.3449430676490288, "grad_norm": 2.414275575818596, "learning_rate": 4.197726139019792e-06, "loss": 1.3093, "step": 2008 }, { "epoch": 1.3456128600133959, "grad_norm": 2.82243324997923, "learning_rate": 4.196653044494107e-06, "loss": 1.3587, "step": 2009 }, { "epoch": 1.346282652377763, "grad_norm": 2.282916700783431, "learning_rate": 4.195579370142736e-06, "loss": 1.2648, "step": 2010 }, { "epoch": 1.34695244474213, "grad_norm": 2.30528977563551, "learning_rate": 4.194505116332606e-06, "loss": 1.3463, "step": 2011 }, { "epoch": 1.347622237106497, "grad_norm": 2.5400088935141096, "learning_rate": 4.193430283430837e-06, "loss": 1.4523, "step": 2012 }, { "epoch": 1.348292029470864, "grad_norm": 3.1969034658120643, "learning_rate": 4.192354871804753e-06, "loss": 1.3931, "step": 2013 }, { "epoch": 1.348961821835231, "grad_norm": 2.483546372392077, "learning_rate": 4.19127888182187e-06, "loss": 1.0659, "step": 2014 }, { "epoch": 1.349631614199598, "grad_norm": 3.234563597672345, "learning_rate": 4.190202313849905e-06, "loss": 1.3043, "step": 2015 }, { "epoch": 1.3503014065639651, "grad_norm": 7.648547089477415, "learning_rate": 4.189125168256771e-06, "loss": 1.2908, "step": 2016 }, { "epoch": 1.3509711989283322, "grad_norm": 2.888523582368992, "learning_rate": 4.18804744541058e-06, "loss": 1.3257, "step": 2017 }, { "epoch": 1.3516409912926992, "grad_norm": 2.347748819319756, "learning_rate": 4.18696914567964e-06, "loss": 1.4503, "step": 2018 }, { "epoch": 1.3523107836570663, "grad_norm": 2.3744923924471437, "learning_rate": 4.185890269432456e-06, "loss": 1.5029, "step": 2019 }, { "epoch": 1.3529805760214333, "grad_norm": 2.4522915863021706, "learning_rate": 4.184810817037731e-06, "loss": 1.4823, "step": 2020 }, { "epoch": 1.3536503683858003, "grad_norm": 2.5357826097121503, "learning_rate": 4.183730788864365e-06, "loss": 1.4844, "step": 2021 }, { "epoch": 1.3543201607501674, "grad_norm": 2.550062739211584, "learning_rate": 4.182650185281452e-06, "loss": 1.4762, "step": 2022 }, { "epoch": 1.3549899531145346, "grad_norm": 2.6260548619474524, "learning_rate": 4.181569006658286e-06, "loss": 1.4672, "step": 2023 }, { "epoch": 1.3556597454789014, "grad_norm": 2.5754882259619953, "learning_rate": 4.1804872533643574e-06, "loss": 1.2573, "step": 2024 }, { "epoch": 1.3563295378432687, "grad_norm": 2.4500982332919783, "learning_rate": 4.17940492576935e-06, "loss": 1.2399, "step": 2025 }, { "epoch": 1.3569993302076355, "grad_norm": 2.23666452110769, "learning_rate": 4.1783220242431465e-06, "loss": 1.2268, "step": 2026 }, { "epoch": 1.3576691225720028, "grad_norm": 2.363111246566829, "learning_rate": 4.177238549155825e-06, "loss": 1.3468, "step": 2027 }, { "epoch": 1.3583389149363696, "grad_norm": 2.216099491993554, "learning_rate": 4.17615450087766e-06, "loss": 1.2486, "step": 2028 }, { "epoch": 1.3590087073007369, "grad_norm": 2.5998652835474525, "learning_rate": 4.175069879779121e-06, "loss": 1.2636, "step": 2029 }, { "epoch": 1.3596784996651037, "grad_norm": 3.0452924721367074, "learning_rate": 4.173984686230873e-06, "loss": 1.4446, "step": 2030 }, { "epoch": 1.360348292029471, "grad_norm": 4.739982648065401, "learning_rate": 4.172898920603779e-06, "loss": 1.4015, "step": 2031 }, { "epoch": 1.361018084393838, "grad_norm": 3.4524911344807565, "learning_rate": 4.171812583268896e-06, "loss": 1.138, "step": 2032 }, { "epoch": 1.361687876758205, "grad_norm": 3.1001320538168295, "learning_rate": 4.170725674597474e-06, "loss": 1.2369, "step": 2033 }, { "epoch": 1.362357669122572, "grad_norm": 2.8054244852244743, "learning_rate": 4.169638194960962e-06, "loss": 1.1898, "step": 2034 }, { "epoch": 1.363027461486939, "grad_norm": 2.212718115210865, "learning_rate": 4.168550144731004e-06, "loss": 1.4147, "step": 2035 }, { "epoch": 1.3636972538513061, "grad_norm": 3.4546930941763203, "learning_rate": 4.1674615242794346e-06, "loss": 1.4309, "step": 2036 }, { "epoch": 1.3643670462156732, "grad_norm": 2.5245221177940365, "learning_rate": 4.166372333978289e-06, "loss": 1.3728, "step": 2037 }, { "epoch": 1.3650368385800402, "grad_norm": 2.6113751933590645, "learning_rate": 4.165282574199793e-06, "loss": 1.2459, "step": 2038 }, { "epoch": 1.3657066309444073, "grad_norm": 2.3323821183656857, "learning_rate": 4.164192245316369e-06, "loss": 1.3683, "step": 2039 }, { "epoch": 1.3663764233087743, "grad_norm": 2.2001982474464157, "learning_rate": 4.163101347700633e-06, "loss": 1.4222, "step": 2040 }, { "epoch": 1.3670462156731413, "grad_norm": 3.412453918061161, "learning_rate": 4.162009881725396e-06, "loss": 1.1731, "step": 2041 }, { "epoch": 1.3677160080375084, "grad_norm": 2.481666436913283, "learning_rate": 4.160917847763663e-06, "loss": 1.5198, "step": 2042 }, { "epoch": 1.3683858004018754, "grad_norm": 2.2416245194811757, "learning_rate": 4.159825246188633e-06, "loss": 1.3169, "step": 2043 }, { "epoch": 1.3690555927662424, "grad_norm": 2.9227854053021867, "learning_rate": 4.1587320773737e-06, "loss": 1.391, "step": 2044 }, { "epoch": 1.3697253851306095, "grad_norm": 2.840038446411312, "learning_rate": 4.1576383416924485e-06, "loss": 1.4218, "step": 2045 }, { "epoch": 1.3703951774949765, "grad_norm": 2.1987307795676903, "learning_rate": 4.156544039518662e-06, "loss": 1.1897, "step": 2046 }, { "epoch": 1.3710649698593436, "grad_norm": 2.1800026283205938, "learning_rate": 4.155449171226312e-06, "loss": 1.2775, "step": 2047 }, { "epoch": 1.3717347622237106, "grad_norm": 2.5800858435615224, "learning_rate": 4.154353737189568e-06, "loss": 1.4474, "step": 2048 }, { "epoch": 1.3724045545880776, "grad_norm": 2.1785657478115676, "learning_rate": 4.153257737782791e-06, "loss": 1.4173, "step": 2049 }, { "epoch": 1.3730743469524447, "grad_norm": 2.778239784859261, "learning_rate": 4.152161173380534e-06, "loss": 1.4569, "step": 2050 }, { "epoch": 1.3737441393168117, "grad_norm": 3.332289762071735, "learning_rate": 4.151064044357545e-06, "loss": 1.1242, "step": 2051 }, { "epoch": 1.3744139316811788, "grad_norm": 3.0152455805690757, "learning_rate": 4.149966351088765e-06, "loss": 1.3539, "step": 2052 }, { "epoch": 1.3750837240455458, "grad_norm": 3.863428302715833, "learning_rate": 4.148868093949326e-06, "loss": 1.4196, "step": 2053 }, { "epoch": 1.375753516409913, "grad_norm": 3.0270638447218343, "learning_rate": 4.147769273314555e-06, "loss": 1.322, "step": 2054 }, { "epoch": 1.3764233087742799, "grad_norm": 2.746722974597434, "learning_rate": 4.146669889559969e-06, "loss": 1.3862, "step": 2055 }, { "epoch": 1.3770931011386471, "grad_norm": 2.14464633641333, "learning_rate": 4.145569943061281e-06, "loss": 1.2731, "step": 2056 }, { "epoch": 1.377762893503014, "grad_norm": 2.4845390332633928, "learning_rate": 4.144469434194391e-06, "loss": 1.4439, "step": 2057 }, { "epoch": 1.3784326858673812, "grad_norm": 2.385627063228069, "learning_rate": 4.143368363335397e-06, "loss": 1.5009, "step": 2058 }, { "epoch": 1.379102478231748, "grad_norm": 2.6273303682135416, "learning_rate": 4.142266730860586e-06, "loss": 1.3967, "step": 2059 }, { "epoch": 1.3797722705961153, "grad_norm": 3.480231378413318, "learning_rate": 4.141164537146436e-06, "loss": 1.2798, "step": 2060 }, { "epoch": 1.3804420629604823, "grad_norm": 3.9572764641238103, "learning_rate": 4.140061782569619e-06, "loss": 1.4108, "step": 2061 }, { "epoch": 1.3811118553248494, "grad_norm": 2.6655907113520043, "learning_rate": 4.138958467506998e-06, "loss": 1.4746, "step": 2062 }, { "epoch": 1.3817816476892164, "grad_norm": 2.379334689029434, "learning_rate": 4.137854592335627e-06, "loss": 1.3998, "step": 2063 }, { "epoch": 1.3824514400535834, "grad_norm": 3.560730585033446, "learning_rate": 4.136750157432752e-06, "loss": 1.1769, "step": 2064 }, { "epoch": 1.3831212324179505, "grad_norm": 3.299380143422073, "learning_rate": 4.13564516317581e-06, "loss": 1.4087, "step": 2065 }, { "epoch": 1.3837910247823175, "grad_norm": 5.623137806907439, "learning_rate": 4.134539609942428e-06, "loss": 1.3528, "step": 2066 }, { "epoch": 1.3844608171466846, "grad_norm": 2.632187115246899, "learning_rate": 4.133433498110428e-06, "loss": 1.5764, "step": 2067 }, { "epoch": 1.3851306095110516, "grad_norm": 2.6545795417453886, "learning_rate": 4.132326828057817e-06, "loss": 1.4782, "step": 2068 }, { "epoch": 1.3858004018754186, "grad_norm": 2.2786897128127714, "learning_rate": 4.131219600162799e-06, "loss": 1.2198, "step": 2069 }, { "epoch": 1.3864701942397857, "grad_norm": 3.0824868627320967, "learning_rate": 4.130111814803763e-06, "loss": 1.3908, "step": 2070 }, { "epoch": 1.3871399866041527, "grad_norm": 3.345604526116789, "learning_rate": 4.129003472359292e-06, "loss": 1.4268, "step": 2071 }, { "epoch": 1.3878097789685198, "grad_norm": 2.575937788432137, "learning_rate": 4.127894573208159e-06, "loss": 1.5306, "step": 2072 }, { "epoch": 1.3884795713328868, "grad_norm": 2.3598077509717355, "learning_rate": 4.126785117729326e-06, "loss": 1.4984, "step": 2073 }, { "epoch": 1.3891493636972538, "grad_norm": 3.826480924294857, "learning_rate": 4.1256751063019455e-06, "loss": 1.3553, "step": 2074 }, { "epoch": 1.3898191560616209, "grad_norm": 2.604077900047402, "learning_rate": 4.124564539305361e-06, "loss": 1.5107, "step": 2075 }, { "epoch": 1.390488948425988, "grad_norm": 3.51389206093538, "learning_rate": 4.123453417119105e-06, "loss": 1.1871, "step": 2076 }, { "epoch": 1.391158740790355, "grad_norm": 5.000938304399281, "learning_rate": 4.122341740122901e-06, "loss": 1.1025, "step": 2077 }, { "epoch": 1.391828533154722, "grad_norm": 2.510878088515939, "learning_rate": 4.121229508696658e-06, "loss": 1.44, "step": 2078 }, { "epoch": 1.392498325519089, "grad_norm": 2.7823546017234553, "learning_rate": 4.1201167232204805e-06, "loss": 1.1819, "step": 2079 }, { "epoch": 1.393168117883456, "grad_norm": 2.9898447151983905, "learning_rate": 4.119003384074658e-06, "loss": 1.3055, "step": 2080 }, { "epoch": 1.393837910247823, "grad_norm": 3.1827145669754358, "learning_rate": 4.117889491639671e-06, "loss": 1.1177, "step": 2081 }, { "epoch": 1.3945077026121901, "grad_norm": 3.8933922712506015, "learning_rate": 4.1167750462961875e-06, "loss": 1.3404, "step": 2082 }, { "epoch": 1.3951774949765572, "grad_norm": 2.589100594162154, "learning_rate": 4.1156600484250676e-06, "loss": 1.4276, "step": 2083 }, { "epoch": 1.3958472873409242, "grad_norm": 2.29940992985469, "learning_rate": 4.114544498407356e-06, "loss": 1.3138, "step": 2084 }, { "epoch": 1.3965170797052915, "grad_norm": 2.345861022742436, "learning_rate": 4.113428396624291e-06, "loss": 1.4523, "step": 2085 }, { "epoch": 1.3971868720696583, "grad_norm": 2.23995462048332, "learning_rate": 4.112311743457293e-06, "loss": 1.2939, "step": 2086 }, { "epoch": 1.3978566644340256, "grad_norm": 2.6093716065551074, "learning_rate": 4.111194539287978e-06, "loss": 1.3674, "step": 2087 }, { "epoch": 1.3985264567983924, "grad_norm": 2.3777250466291946, "learning_rate": 4.110076784498146e-06, "loss": 1.4255, "step": 2088 }, { "epoch": 1.3991962491627596, "grad_norm": 2.3217358878971184, "learning_rate": 4.1089584794697854e-06, "loss": 1.2724, "step": 2089 }, { "epoch": 1.3998660415271265, "grad_norm": 3.5230998834045724, "learning_rate": 4.107839624585073e-06, "loss": 1.3512, "step": 2090 }, { "epoch": 1.4005358338914937, "grad_norm": 2.3158172521998868, "learning_rate": 4.1067202202263756e-06, "loss": 1.2487, "step": 2091 }, { "epoch": 1.4012056262558608, "grad_norm": 2.7870060333398277, "learning_rate": 4.105600266776245e-06, "loss": 1.4236, "step": 2092 }, { "epoch": 1.4018754186202278, "grad_norm": 2.4870925700234574, "learning_rate": 4.10447976461742e-06, "loss": 1.4073, "step": 2093 }, { "epoch": 1.4025452109845948, "grad_norm": 2.269493591067875, "learning_rate": 4.103358714132832e-06, "loss": 1.4296, "step": 2094 }, { "epoch": 1.4032150033489619, "grad_norm": 3.178099220627257, "learning_rate": 4.102237115705594e-06, "loss": 1.3335, "step": 2095 }, { "epoch": 1.403884795713329, "grad_norm": 2.1960985914493554, "learning_rate": 4.101114969719009e-06, "loss": 1.2368, "step": 2096 }, { "epoch": 1.404554588077696, "grad_norm": 4.536266545270564, "learning_rate": 4.099992276556567e-06, "loss": 1.1723, "step": 2097 }, { "epoch": 1.405224380442063, "grad_norm": 4.183763817622146, "learning_rate": 4.0988690366019446e-06, "loss": 1.3696, "step": 2098 }, { "epoch": 1.40589417280643, "grad_norm": 4.001191913602328, "learning_rate": 4.097745250239006e-06, "loss": 1.3866, "step": 2099 }, { "epoch": 1.406563965170797, "grad_norm": 2.311414605388674, "learning_rate": 4.0966209178518e-06, "loss": 1.4132, "step": 2100 }, { "epoch": 1.407233757535164, "grad_norm": 8.426691144033818, "learning_rate": 4.095496039824566e-06, "loss": 1.3854, "step": 2101 }, { "epoch": 1.4079035498995311, "grad_norm": 4.624001765763422, "learning_rate": 4.094370616541725e-06, "loss": 1.3147, "step": 2102 }, { "epoch": 1.4085733422638982, "grad_norm": 2.237838481642395, "learning_rate": 4.093244648387889e-06, "loss": 1.4372, "step": 2103 }, { "epoch": 1.4092431346282652, "grad_norm": 2.63042408752709, "learning_rate": 4.092118135747852e-06, "loss": 1.354, "step": 2104 }, { "epoch": 1.4099129269926323, "grad_norm": 2.3243871010807555, "learning_rate": 4.0909910790065965e-06, "loss": 1.4646, "step": 2105 }, { "epoch": 1.4105827193569993, "grad_norm": 3.328198732300638, "learning_rate": 4.089863478549291e-06, "loss": 1.4193, "step": 2106 }, { "epoch": 1.4112525117213663, "grad_norm": 3.249300743999498, "learning_rate": 4.088735334761289e-06, "loss": 1.2184, "step": 2107 }, { "epoch": 1.4119223040857334, "grad_norm": 2.255412878765539, "learning_rate": 4.087606648028131e-06, "loss": 1.499, "step": 2108 }, { "epoch": 1.4125920964501004, "grad_norm": 3.3069568256483692, "learning_rate": 4.08647741873554e-06, "loss": 1.5108, "step": 2109 }, { "epoch": 1.4132618888144675, "grad_norm": 3.1034446945015066, "learning_rate": 4.085347647269429e-06, "loss": 1.391, "step": 2110 }, { "epoch": 1.4139316811788345, "grad_norm": 2.3797266158189947, "learning_rate": 4.0842173340158905e-06, "loss": 1.3106, "step": 2111 }, { "epoch": 1.4146014735432015, "grad_norm": 2.9646515417075365, "learning_rate": 4.083086479361209e-06, "loss": 1.3065, "step": 2112 }, { "epoch": 1.4152712659075686, "grad_norm": 2.292392292464056, "learning_rate": 4.081955083691848e-06, "loss": 1.39, "step": 2113 }, { "epoch": 1.4159410582719358, "grad_norm": 2.1096926099945206, "learning_rate": 4.080823147394459e-06, "loss": 1.4129, "step": 2114 }, { "epoch": 1.4166108506363027, "grad_norm": 2.19954044150368, "learning_rate": 4.079690670855877e-06, "loss": 1.2331, "step": 2115 }, { "epoch": 1.41728064300067, "grad_norm": 2.7318104368462963, "learning_rate": 4.078557654463123e-06, "loss": 1.5646, "step": 2116 }, { "epoch": 1.4179504353650367, "grad_norm": 2.272255880448683, "learning_rate": 4.077424098603402e-06, "loss": 1.4575, "step": 2117 }, { "epoch": 1.418620227729404, "grad_norm": 2.234390826357997, "learning_rate": 4.0762900036641015e-06, "loss": 1.323, "step": 2118 }, { "epoch": 1.4192900200937708, "grad_norm": 4.007350170859394, "learning_rate": 4.075155370032796e-06, "loss": 1.294, "step": 2119 }, { "epoch": 1.419959812458138, "grad_norm": 3.2097273225311507, "learning_rate": 4.074020198097244e-06, "loss": 1.1858, "step": 2120 }, { "epoch": 1.4206296048225049, "grad_norm": 2.2546947796710155, "learning_rate": 4.072884488245384e-06, "loss": 1.1944, "step": 2121 }, { "epoch": 1.4212993971868721, "grad_norm": 2.310458366105694, "learning_rate": 4.071748240865343e-06, "loss": 1.5145, "step": 2122 }, { "epoch": 1.4219691895512392, "grad_norm": 2.8209859812937528, "learning_rate": 4.070611456345429e-06, "loss": 1.2121, "step": 2123 }, { "epoch": 1.4226389819156062, "grad_norm": 2.360372221795341, "learning_rate": 4.069474135074135e-06, "loss": 1.3265, "step": 2124 }, { "epoch": 1.4233087742799733, "grad_norm": 3.9147723671792245, "learning_rate": 4.068336277440137e-06, "loss": 1.2196, "step": 2125 }, { "epoch": 1.4239785666443403, "grad_norm": 2.5818652065354035, "learning_rate": 4.067197883832294e-06, "loss": 1.5831, "step": 2126 }, { "epoch": 1.4246483590087073, "grad_norm": 3.0941731102382284, "learning_rate": 4.066058954639648e-06, "loss": 1.3805, "step": 2127 }, { "epoch": 1.4253181513730744, "grad_norm": 2.438463488552508, "learning_rate": 4.064919490251425e-06, "loss": 1.4383, "step": 2128 }, { "epoch": 1.4259879437374414, "grad_norm": 2.4549056305872963, "learning_rate": 4.063779491057033e-06, "loss": 1.3322, "step": 2129 }, { "epoch": 1.4266577361018085, "grad_norm": 3.227498314289574, "learning_rate": 4.062638957446062e-06, "loss": 1.3538, "step": 2130 }, { "epoch": 1.4273275284661755, "grad_norm": 5.954527693283428, "learning_rate": 4.0614978898082865e-06, "loss": 1.1507, "step": 2131 }, { "epoch": 1.4279973208305425, "grad_norm": 2.6481114652929003, "learning_rate": 4.060356288533664e-06, "loss": 1.4784, "step": 2132 }, { "epoch": 1.4286671131949096, "grad_norm": 2.4028903576611644, "learning_rate": 4.059214154012331e-06, "loss": 1.4902, "step": 2133 }, { "epoch": 1.4293369055592766, "grad_norm": 2.443201246036188, "learning_rate": 4.058071486634609e-06, "loss": 1.3378, "step": 2134 }, { "epoch": 1.4300066979236437, "grad_norm": 4.457088381232623, "learning_rate": 4.056928286791001e-06, "loss": 1.2893, "step": 2135 }, { "epoch": 1.4306764902880107, "grad_norm": 2.1614388028583464, "learning_rate": 4.055784554872192e-06, "loss": 1.2856, "step": 2136 }, { "epoch": 1.4313462826523777, "grad_norm": 2.3197772619444503, "learning_rate": 4.054640291269049e-06, "loss": 1.4514, "step": 2137 }, { "epoch": 1.4320160750167448, "grad_norm": 2.952850118562335, "learning_rate": 4.053495496372619e-06, "loss": 1.1703, "step": 2138 }, { "epoch": 1.4326858673811118, "grad_norm": 2.2385693659874577, "learning_rate": 4.052350170574132e-06, "loss": 1.3504, "step": 2139 }, { "epoch": 1.4333556597454788, "grad_norm": 2.3752722368663295, "learning_rate": 4.0512043142650015e-06, "loss": 1.3123, "step": 2140 }, { "epoch": 1.4340254521098459, "grad_norm": 2.1808757278806263, "learning_rate": 4.050057927836819e-06, "loss": 1.1907, "step": 2141 }, { "epoch": 1.434695244474213, "grad_norm": 4.127738787267381, "learning_rate": 4.048911011681357e-06, "loss": 1.2545, "step": 2142 }, { "epoch": 1.43536503683858, "grad_norm": 2.7135856283490245, "learning_rate": 4.047763566190573e-06, "loss": 1.4625, "step": 2143 }, { "epoch": 1.436034829202947, "grad_norm": 3.6249393242337913, "learning_rate": 4.0466155917566e-06, "loss": 1.2482, "step": 2144 }, { "epoch": 1.4367046215673143, "grad_norm": 2.5116236621150634, "learning_rate": 4.045467088771758e-06, "loss": 1.4411, "step": 2145 }, { "epoch": 1.437374413931681, "grad_norm": 2.6472409052077857, "learning_rate": 4.044318057628539e-06, "loss": 1.1199, "step": 2146 }, { "epoch": 1.4380442062960483, "grad_norm": 3.66327500562587, "learning_rate": 4.043168498719626e-06, "loss": 1.3179, "step": 2147 }, { "epoch": 1.4387139986604152, "grad_norm": 2.2964470596530275, "learning_rate": 4.042018412437874e-06, "loss": 1.3604, "step": 2148 }, { "epoch": 1.4393837910247824, "grad_norm": 2.422242392034623, "learning_rate": 4.040867799176323e-06, "loss": 1.4206, "step": 2149 }, { "epoch": 1.4400535833891492, "grad_norm": 2.456864548600033, "learning_rate": 4.03971665932819e-06, "loss": 1.3636, "step": 2150 }, { "epoch": 1.4407233757535165, "grad_norm": 2.171684650715512, "learning_rate": 4.038564993286873e-06, "loss": 1.3365, "step": 2151 }, { "epoch": 1.4413931681178835, "grad_norm": 3.1245804290160772, "learning_rate": 4.037412801445952e-06, "loss": 1.3494, "step": 2152 }, { "epoch": 1.4420629604822506, "grad_norm": 2.881356963415187, "learning_rate": 4.036260084199183e-06, "loss": 1.4574, "step": 2153 }, { "epoch": 1.4427327528466176, "grad_norm": 2.392058081402394, "learning_rate": 4.035106841940503e-06, "loss": 1.5324, "step": 2154 }, { "epoch": 1.4434025452109847, "grad_norm": 3.7708500770356492, "learning_rate": 4.033953075064032e-06, "loss": 1.2559, "step": 2155 }, { "epoch": 1.4440723375753517, "grad_norm": 4.843667232093005, "learning_rate": 4.032798783964063e-06, "loss": 1.2505, "step": 2156 }, { "epoch": 1.4447421299397187, "grad_norm": 2.503860092009062, "learning_rate": 4.0316439690350725e-06, "loss": 1.5272, "step": 2157 }, { "epoch": 1.4454119223040858, "grad_norm": 2.3325122842565387, "learning_rate": 4.030488630671714e-06, "loss": 1.2862, "step": 2158 }, { "epoch": 1.4460817146684528, "grad_norm": 2.2008767031926357, "learning_rate": 4.029332769268822e-06, "loss": 1.4239, "step": 2159 }, { "epoch": 1.4467515070328199, "grad_norm": 3.9135721864175808, "learning_rate": 4.028176385221408e-06, "loss": 1.3147, "step": 2160 }, { "epoch": 1.447421299397187, "grad_norm": 2.2997890734688484, "learning_rate": 4.02701947892466e-06, "loss": 1.5431, "step": 2161 }, { "epoch": 1.448091091761554, "grad_norm": 5.128796140967415, "learning_rate": 4.02586205077395e-06, "loss": 1.2073, "step": 2162 }, { "epoch": 1.448760884125921, "grad_norm": 2.954036402902186, "learning_rate": 4.024704101164824e-06, "loss": 1.3507, "step": 2163 }, { "epoch": 1.449430676490288, "grad_norm": 2.258258117701872, "learning_rate": 4.023545630493009e-06, "loss": 1.3843, "step": 2164 }, { "epoch": 1.450100468854655, "grad_norm": 3.3822829072704206, "learning_rate": 4.0223866391544075e-06, "loss": 1.43, "step": 2165 }, { "epoch": 1.450770261219022, "grad_norm": 2.388307794964528, "learning_rate": 4.0212271275451e-06, "loss": 1.2929, "step": 2166 }, { "epoch": 1.4514400535833891, "grad_norm": 2.3176628193482527, "learning_rate": 4.020067096061347e-06, "loss": 1.3543, "step": 2167 }, { "epoch": 1.4521098459477562, "grad_norm": 2.299714888519748, "learning_rate": 4.018906545099587e-06, "loss": 1.2584, "step": 2168 }, { "epoch": 1.4527796383121232, "grad_norm": 2.296061488100173, "learning_rate": 4.017745475056431e-06, "loss": 1.4566, "step": 2169 }, { "epoch": 1.4534494306764902, "grad_norm": 2.2690550309754296, "learning_rate": 4.016583886328673e-06, "loss": 1.3149, "step": 2170 }, { "epoch": 1.4541192230408573, "grad_norm": 2.320877380104597, "learning_rate": 4.015421779313283e-06, "loss": 1.3364, "step": 2171 }, { "epoch": 1.4547890154052243, "grad_norm": 2.3233740489695456, "learning_rate": 4.014259154407405e-06, "loss": 1.145, "step": 2172 }, { "epoch": 1.4554588077695914, "grad_norm": 2.241617906199806, "learning_rate": 4.013096012008365e-06, "loss": 1.4469, "step": 2173 }, { "epoch": 1.4561286001339584, "grad_norm": 2.7575231886989937, "learning_rate": 4.011932352513661e-06, "loss": 1.3702, "step": 2174 }, { "epoch": 1.4567983924983254, "grad_norm": 2.630775632183631, "learning_rate": 4.010768176320971e-06, "loss": 1.3122, "step": 2175 }, { "epoch": 1.4574681848626927, "grad_norm": 8.11328519180718, "learning_rate": 4.009603483828147e-06, "loss": 1.3232, "step": 2176 }, { "epoch": 1.4581379772270595, "grad_norm": 2.285036271288875, "learning_rate": 4.008438275433221e-06, "loss": 1.4211, "step": 2177 }, { "epoch": 1.4588077695914268, "grad_norm": 2.596974578678349, "learning_rate": 4.007272551534398e-06, "loss": 1.3306, "step": 2178 }, { "epoch": 1.4594775619557936, "grad_norm": 14.170845017200774, "learning_rate": 4.006106312530061e-06, "loss": 1.4641, "step": 2179 }, { "epoch": 1.4601473543201609, "grad_norm": 5.12298277078257, "learning_rate": 4.0049395588187676e-06, "loss": 1.2047, "step": 2180 }, { "epoch": 1.4608171466845277, "grad_norm": 5.082307303144062, "learning_rate": 4.003772290799252e-06, "loss": 1.2571, "step": 2181 }, { "epoch": 1.461486939048895, "grad_norm": 6.387762714165787, "learning_rate": 4.002604508870425e-06, "loss": 1.353, "step": 2182 }, { "epoch": 1.462156731413262, "grad_norm": 2.1743138211048336, "learning_rate": 4.001436213431372e-06, "loss": 1.2561, "step": 2183 }, { "epoch": 1.462826523777629, "grad_norm": 2.428207034005932, "learning_rate": 4.000267404881354e-06, "loss": 1.3853, "step": 2184 }, { "epoch": 1.463496316141996, "grad_norm": 2.375759639458155, "learning_rate": 3.999098083619808e-06, "loss": 1.3236, "step": 2185 }, { "epoch": 1.464166108506363, "grad_norm": 2.206296293194174, "learning_rate": 3.9979282500463455e-06, "loss": 1.3191, "step": 2186 }, { "epoch": 1.4648359008707301, "grad_norm": 2.810068978551643, "learning_rate": 3.996757904560754e-06, "loss": 1.2351, "step": 2187 }, { "epoch": 1.4655056932350972, "grad_norm": 2.371163782825923, "learning_rate": 3.995587047562994e-06, "loss": 1.4953, "step": 2188 }, { "epoch": 1.4661754855994642, "grad_norm": 2.273566155326284, "learning_rate": 3.994415679453202e-06, "loss": 1.3746, "step": 2189 }, { "epoch": 1.4668452779638312, "grad_norm": 2.192724916402288, "learning_rate": 3.993243800631691e-06, "loss": 1.3746, "step": 2190 }, { "epoch": 1.4675150703281983, "grad_norm": 2.5491530532809454, "learning_rate": 3.992071411498945e-06, "loss": 1.3905, "step": 2191 }, { "epoch": 1.4681848626925653, "grad_norm": 3.613937109878744, "learning_rate": 3.990898512455625e-06, "loss": 1.3242, "step": 2192 }, { "epoch": 1.4688546550569324, "grad_norm": 2.885235909800943, "learning_rate": 3.9897251039025655e-06, "loss": 1.2104, "step": 2193 }, { "epoch": 1.4695244474212994, "grad_norm": 6.561048104586148, "learning_rate": 3.988551186240775e-06, "loss": 1.3747, "step": 2194 }, { "epoch": 1.4701942397856664, "grad_norm": 3.072857275835819, "learning_rate": 3.987376759871436e-06, "loss": 1.3457, "step": 2195 }, { "epoch": 1.4708640321500335, "grad_norm": 4.923165326005953, "learning_rate": 3.986201825195905e-06, "loss": 1.4192, "step": 2196 }, { "epoch": 1.4715338245144005, "grad_norm": 2.435043424154647, "learning_rate": 3.985026382615712e-06, "loss": 1.2662, "step": 2197 }, { "epoch": 1.4722036168787676, "grad_norm": 2.17530770191589, "learning_rate": 3.983850432532561e-06, "loss": 1.3543, "step": 2198 }, { "epoch": 1.4728734092431346, "grad_norm": 2.2251846767488486, "learning_rate": 3.982673975348328e-06, "loss": 1.2578, "step": 2199 }, { "epoch": 1.4735432016075016, "grad_norm": 3.315669536124128, "learning_rate": 3.981497011465065e-06, "loss": 1.287, "step": 2200 }, { "epoch": 1.4742129939718687, "grad_norm": 2.751242019759212, "learning_rate": 3.980319541284995e-06, "loss": 1.1782, "step": 2201 }, { "epoch": 1.4748827863362357, "grad_norm": 2.5117156860804233, "learning_rate": 3.979141565210514e-06, "loss": 1.3825, "step": 2202 }, { "epoch": 1.4755525787006027, "grad_norm": 2.6187705637947034, "learning_rate": 3.977963083644192e-06, "loss": 1.3786, "step": 2203 }, { "epoch": 1.4762223710649698, "grad_norm": 3.1281871962690193, "learning_rate": 3.9767840969887726e-06, "loss": 1.3171, "step": 2204 }, { "epoch": 1.4768921634293368, "grad_norm": 2.2312784460293003, "learning_rate": 3.975604605647167e-06, "loss": 1.3573, "step": 2205 }, { "epoch": 1.4775619557937039, "grad_norm": 2.131817287494056, "learning_rate": 3.974424610022467e-06, "loss": 1.3148, "step": 2206 }, { "epoch": 1.4782317481580711, "grad_norm": 3.021377194076534, "learning_rate": 3.97324411051793e-06, "loss": 1.3312, "step": 2207 }, { "epoch": 1.478901540522438, "grad_norm": 2.1685706618884293, "learning_rate": 3.972063107536987e-06, "loss": 1.2804, "step": 2208 }, { "epoch": 1.4795713328868052, "grad_norm": 2.6723235132500407, "learning_rate": 3.970881601483244e-06, "loss": 1.2354, "step": 2209 }, { "epoch": 1.480241125251172, "grad_norm": 2.4446033923455053, "learning_rate": 3.969699592760476e-06, "loss": 1.3557, "step": 2210 }, { "epoch": 1.4809109176155393, "grad_norm": 2.5590356111126162, "learning_rate": 3.968517081772629e-06, "loss": 1.4075, "step": 2211 }, { "epoch": 1.481580709979906, "grad_norm": 3.2513582309129987, "learning_rate": 3.967334068923825e-06, "loss": 1.518, "step": 2212 }, { "epoch": 1.4822505023442734, "grad_norm": 2.800660208959382, "learning_rate": 3.966150554618355e-06, "loss": 1.5111, "step": 2213 }, { "epoch": 1.4829202947086404, "grad_norm": 2.4311779796320954, "learning_rate": 3.964966539260677e-06, "loss": 1.2838, "step": 2214 }, { "epoch": 1.4835900870730074, "grad_norm": 2.675666492157799, "learning_rate": 3.963782023255429e-06, "loss": 1.3552, "step": 2215 }, { "epoch": 1.4842598794373745, "grad_norm": 2.6350192459645836, "learning_rate": 3.962597007007414e-06, "loss": 1.4188, "step": 2216 }, { "epoch": 1.4849296718017415, "grad_norm": 2.2326777845370285, "learning_rate": 3.961411490921606e-06, "loss": 1.2449, "step": 2217 }, { "epoch": 1.4855994641661086, "grad_norm": 2.5218320994631758, "learning_rate": 3.960225475403153e-06, "loss": 1.418, "step": 2218 }, { "epoch": 1.4862692565304756, "grad_norm": 2.1892436200985084, "learning_rate": 3.959038960857372e-06, "loss": 1.279, "step": 2219 }, { "epoch": 1.4869390488948426, "grad_norm": 2.1245935484383884, "learning_rate": 3.95785194768975e-06, "loss": 1.3124, "step": 2220 }, { "epoch": 1.4876088412592097, "grad_norm": 3.8255728707540966, "learning_rate": 3.956664436305945e-06, "loss": 1.3594, "step": 2221 }, { "epoch": 1.4882786336235767, "grad_norm": 2.293738084041619, "learning_rate": 3.9554764271117844e-06, "loss": 1.2902, "step": 2222 }, { "epoch": 1.4889484259879437, "grad_norm": 2.6686161647072533, "learning_rate": 3.954287920513267e-06, "loss": 1.4266, "step": 2223 }, { "epoch": 1.4896182183523108, "grad_norm": 3.093744102376783, "learning_rate": 3.953098916916563e-06, "loss": 1.3543, "step": 2224 }, { "epoch": 1.4902880107166778, "grad_norm": 2.369917101575235, "learning_rate": 3.951909416728007e-06, "loss": 1.3853, "step": 2225 }, { "epoch": 1.4909578030810449, "grad_norm": 2.388321377695414, "learning_rate": 3.950719420354109e-06, "loss": 1.4621, "step": 2226 }, { "epoch": 1.491627595445412, "grad_norm": 2.352828686867738, "learning_rate": 3.9495289282015455e-06, "loss": 1.2699, "step": 2227 }, { "epoch": 1.492297387809779, "grad_norm": 2.324707522717863, "learning_rate": 3.948337940677165e-06, "loss": 1.4496, "step": 2228 }, { "epoch": 1.492967180174146, "grad_norm": 2.1612125860303726, "learning_rate": 3.947146458187982e-06, "loss": 1.4526, "step": 2229 }, { "epoch": 1.493636972538513, "grad_norm": 2.8134921735919254, "learning_rate": 3.945954481141181e-06, "loss": 1.2561, "step": 2230 }, { "epoch": 1.49430676490288, "grad_norm": 2.761817357589984, "learning_rate": 3.9447620099441195e-06, "loss": 1.3522, "step": 2231 }, { "epoch": 1.494976557267247, "grad_norm": 2.6343569018174, "learning_rate": 3.943569045004317e-06, "loss": 1.1243, "step": 2232 }, { "epoch": 1.4956463496316141, "grad_norm": 2.387897918027235, "learning_rate": 3.942375586729469e-06, "loss": 1.5115, "step": 2233 }, { "epoch": 1.4963161419959812, "grad_norm": 2.606111526644013, "learning_rate": 3.941181635527433e-06, "loss": 1.2631, "step": 2234 }, { "epoch": 1.4969859343603482, "grad_norm": 2.4101305078608446, "learning_rate": 3.939987191806239e-06, "loss": 1.3427, "step": 2235 }, { "epoch": 1.4976557267247155, "grad_norm": 3.232287607212448, "learning_rate": 3.938792255974085e-06, "loss": 1.1866, "step": 2236 }, { "epoch": 1.4983255190890823, "grad_norm": 2.9362244219807025, "learning_rate": 3.937596828439336e-06, "loss": 1.3261, "step": 2237 }, { "epoch": 1.4989953114534496, "grad_norm": 2.315249544751344, "learning_rate": 3.936400909610525e-06, "loss": 1.1893, "step": 2238 }, { "epoch": 1.4996651038178164, "grad_norm": 3.9898528717217396, "learning_rate": 3.9352044998963535e-06, "loss": 1.3356, "step": 2239 }, { "epoch": 1.5003348961821836, "grad_norm": 2.956557553013731, "learning_rate": 3.93400759970569e-06, "loss": 1.2732, "step": 2240 }, { "epoch": 1.5010046885465504, "grad_norm": 3.2498198184013485, "learning_rate": 3.932810209447574e-06, "loss": 1.1986, "step": 2241 }, { "epoch": 1.5016744809109177, "grad_norm": 3.3246412468314412, "learning_rate": 3.931612329531207e-06, "loss": 1.3535, "step": 2242 }, { "epoch": 1.5023442732752845, "grad_norm": 2.5529574451545476, "learning_rate": 3.930413960365961e-06, "loss": 1.4094, "step": 2243 }, { "epoch": 1.5030140656396518, "grad_norm": 2.309959730502569, "learning_rate": 3.929215102361376e-06, "loss": 1.4086, "step": 2244 }, { "epoch": 1.5036838580040186, "grad_norm": 2.497412317256827, "learning_rate": 3.928015755927156e-06, "loss": 1.4115, "step": 2245 }, { "epoch": 1.5043536503683859, "grad_norm": 2.4573402516551983, "learning_rate": 3.926815921473176e-06, "loss": 1.1743, "step": 2246 }, { "epoch": 1.5050234427327527, "grad_norm": 2.1476985741125083, "learning_rate": 3.925615599409473e-06, "loss": 1.3291, "step": 2247 }, { "epoch": 1.50569323509712, "grad_norm": 5.5149547145489, "learning_rate": 3.924414790146256e-06, "loss": 1.3773, "step": 2248 }, { "epoch": 1.506363027461487, "grad_norm": 3.3532349483689443, "learning_rate": 3.923213494093896e-06, "loss": 1.327, "step": 2249 }, { "epoch": 1.507032819825854, "grad_norm": 2.3623998351412494, "learning_rate": 3.922011711662932e-06, "loss": 1.323, "step": 2250 }, { "epoch": 1.507702612190221, "grad_norm": 3.6883909113084634, "learning_rate": 3.920809443264069e-06, "loss": 1.261, "step": 2251 }, { "epoch": 1.508372404554588, "grad_norm": 2.2734173368793056, "learning_rate": 3.91960668930818e-06, "loss": 1.4948, "step": 2252 }, { "epoch": 1.5090421969189551, "grad_norm": 2.707056864905993, "learning_rate": 3.918403450206299e-06, "loss": 1.3251, "step": 2253 }, { "epoch": 1.5097119892833222, "grad_norm": 2.4376446138028314, "learning_rate": 3.917199726369634e-06, "loss": 1.5744, "step": 2254 }, { "epoch": 1.5103817816476892, "grad_norm": 2.2586324282487156, "learning_rate": 3.915995518209549e-06, "loss": 1.3823, "step": 2255 }, { "epoch": 1.5110515740120563, "grad_norm": 2.692232696932098, "learning_rate": 3.914790826137581e-06, "loss": 1.3274, "step": 2256 }, { "epoch": 1.5117213663764233, "grad_norm": 4.552305479503411, "learning_rate": 3.913585650565428e-06, "loss": 1.29, "step": 2257 }, { "epoch": 1.5123911587407903, "grad_norm": 2.7777548707775077, "learning_rate": 3.912379991904956e-06, "loss": 1.2283, "step": 2258 }, { "epoch": 1.5130609511051574, "grad_norm": 2.339827129575321, "learning_rate": 3.9111738505681945e-06, "loss": 1.4372, "step": 2259 }, { "epoch": 1.5137307434695244, "grad_norm": 6.361387473929896, "learning_rate": 3.909967226967339e-06, "loss": 1.2411, "step": 2260 }, { "epoch": 1.5144005358338914, "grad_norm": 3.3611302972697903, "learning_rate": 3.908760121514748e-06, "loss": 1.3118, "step": 2261 }, { "epoch": 1.5150703281982585, "grad_norm": 2.410197651850288, "learning_rate": 3.907552534622947e-06, "loss": 1.4239, "step": 2262 }, { "epoch": 1.5157401205626257, "grad_norm": 2.398114970459742, "learning_rate": 3.906344466704624e-06, "loss": 1.2944, "step": 2263 }, { "epoch": 1.5164099129269926, "grad_norm": 3.518304554209832, "learning_rate": 3.905135918172633e-06, "loss": 1.4507, "step": 2264 }, { "epoch": 1.5170797052913598, "grad_norm": 2.3711207393839486, "learning_rate": 3.903926889439991e-06, "loss": 1.3895, "step": 2265 }, { "epoch": 1.5177494976557266, "grad_norm": 2.2100141026714524, "learning_rate": 3.90271738091988e-06, "loss": 1.4648, "step": 2266 }, { "epoch": 1.518419290020094, "grad_norm": 2.984673442040884, "learning_rate": 3.901507393025646e-06, "loss": 1.4253, "step": 2267 }, { "epoch": 1.5190890823844607, "grad_norm": 2.359895367849771, "learning_rate": 3.900296926170798e-06, "loss": 1.3166, "step": 2268 }, { "epoch": 1.519758874748828, "grad_norm": 2.26738688763605, "learning_rate": 3.899085980769009e-06, "loss": 1.3337, "step": 2269 }, { "epoch": 1.5204286671131948, "grad_norm": 2.1951508159047814, "learning_rate": 3.897874557234116e-06, "loss": 1.2811, "step": 2270 }, { "epoch": 1.521098459477562, "grad_norm": 2.2459817022474535, "learning_rate": 3.896662655980119e-06, "loss": 1.2466, "step": 2271 }, { "epoch": 1.5217682518419289, "grad_norm": 4.227749888599309, "learning_rate": 3.895450277421181e-06, "loss": 1.3667, "step": 2272 }, { "epoch": 1.5224380442062961, "grad_norm": 2.2322297890150544, "learning_rate": 3.894237421971628e-06, "loss": 1.248, "step": 2273 }, { "epoch": 1.523107836570663, "grad_norm": 2.2134033832389353, "learning_rate": 3.893024090045951e-06, "loss": 1.1663, "step": 2274 }, { "epoch": 1.5237776289350302, "grad_norm": 8.150455615857686, "learning_rate": 3.8918102820588014e-06, "loss": 1.323, "step": 2275 }, { "epoch": 1.524447421299397, "grad_norm": 3.3239188534739417, "learning_rate": 3.890595998424994e-06, "loss": 1.2694, "step": 2276 }, { "epoch": 1.5251172136637643, "grad_norm": 2.237239280663509, "learning_rate": 3.889381239559506e-06, "loss": 1.5417, "step": 2277 }, { "epoch": 1.5257870060281313, "grad_norm": 2.2698330993749924, "learning_rate": 3.8881660058774776e-06, "loss": 1.2496, "step": 2278 }, { "epoch": 1.5264567983924984, "grad_norm": 2.2259359451071297, "learning_rate": 3.886950297794212e-06, "loss": 1.4139, "step": 2279 }, { "epoch": 1.5271265907568654, "grad_norm": 2.6655833556708206, "learning_rate": 3.885734115725171e-06, "loss": 1.3188, "step": 2280 }, { "epoch": 1.5277963831212324, "grad_norm": 2.8148944085080325, "learning_rate": 3.884517460085983e-06, "loss": 1.2988, "step": 2281 }, { "epoch": 1.5284661754855995, "grad_norm": 2.3298139376049964, "learning_rate": 3.883300331292437e-06, "loss": 1.482, "step": 2282 }, { "epoch": 1.5291359678499665, "grad_norm": 3.7841249202190492, "learning_rate": 3.8820827297604806e-06, "loss": 1.2307, "step": 2283 }, { "epoch": 1.5298057602143336, "grad_norm": 3.9798987667442574, "learning_rate": 3.880864655906226e-06, "loss": 1.2114, "step": 2284 }, { "epoch": 1.5304755525787006, "grad_norm": 2.455161547001405, "learning_rate": 3.879646110145946e-06, "loss": 1.3294, "step": 2285 }, { "epoch": 1.5311453449430676, "grad_norm": 2.5887285134789626, "learning_rate": 3.878427092896074e-06, "loss": 1.5265, "step": 2286 }, { "epoch": 1.5318151373074347, "grad_norm": 2.714830645482012, "learning_rate": 3.877207604573207e-06, "loss": 1.3321, "step": 2287 }, { "epoch": 1.5324849296718017, "grad_norm": 2.8609288659687446, "learning_rate": 3.875987645594099e-06, "loss": 1.1836, "step": 2288 }, { "epoch": 1.5331547220361688, "grad_norm": 2.1533972154433663, "learning_rate": 3.874767216375669e-06, "loss": 1.3989, "step": 2289 }, { "epoch": 1.5338245144005358, "grad_norm": 2.1456078813891857, "learning_rate": 3.873546317334993e-06, "loss": 1.1474, "step": 2290 }, { "epoch": 1.5344943067649028, "grad_norm": 2.19053757460053, "learning_rate": 3.872324948889311e-06, "loss": 1.3336, "step": 2291 }, { "epoch": 1.5351640991292699, "grad_norm": 2.5643825670887375, "learning_rate": 3.8711031114560206e-06, "loss": 1.3228, "step": 2292 }, { "epoch": 1.535833891493637, "grad_norm": 2.305657046916697, "learning_rate": 3.86988080545268e-06, "loss": 1.4118, "step": 2293 }, { "epoch": 1.5365036838580042, "grad_norm": 2.770588301031042, "learning_rate": 3.86865803129701e-06, "loss": 1.3385, "step": 2294 }, { "epoch": 1.537173476222371, "grad_norm": 3.0485946898507783, "learning_rate": 3.8674347894068895e-06, "loss": 1.4467, "step": 2295 }, { "epoch": 1.5378432685867383, "grad_norm": 2.256675580797101, "learning_rate": 3.866211080200356e-06, "loss": 1.2593, "step": 2296 }, { "epoch": 1.538513060951105, "grad_norm": 2.149846590602833, "learning_rate": 3.864986904095612e-06, "loss": 1.3253, "step": 2297 }, { "epoch": 1.5391828533154723, "grad_norm": 2.926260336436944, "learning_rate": 3.8637622615110094e-06, "loss": 1.5187, "step": 2298 }, { "epoch": 1.5398526456798391, "grad_norm": 2.2273182895132777, "learning_rate": 3.8625371528650705e-06, "loss": 1.4099, "step": 2299 }, { "epoch": 1.5405224380442064, "grad_norm": 2.186727947040013, "learning_rate": 3.861311578576471e-06, "loss": 1.364, "step": 2300 }, { "epoch": 1.5411922304085732, "grad_norm": 4.4875995764323395, "learning_rate": 3.860085539064047e-06, "loss": 1.4629, "step": 2301 }, { "epoch": 1.5418620227729405, "grad_norm": 2.2077398931932692, "learning_rate": 3.858859034746793e-06, "loss": 1.3479, "step": 2302 }, { "epoch": 1.5425318151373073, "grad_norm": 3.856080813173084, "learning_rate": 3.857632066043865e-06, "loss": 1.3319, "step": 2303 }, { "epoch": 1.5432016075016746, "grad_norm": 2.2294421025678464, "learning_rate": 3.8564046333745734e-06, "loss": 1.3056, "step": 2304 }, { "epoch": 1.5438713998660414, "grad_norm": 2.514566789963904, "learning_rate": 3.855176737158389e-06, "loss": 1.322, "step": 2305 }, { "epoch": 1.5445411922304086, "grad_norm": 2.0279869430999646, "learning_rate": 3.853948377814943e-06, "loss": 1.2725, "step": 2306 }, { "epoch": 1.5452109845947755, "grad_norm": 2.113362915974821, "learning_rate": 3.8527195557640215e-06, "loss": 1.1801, "step": 2307 }, { "epoch": 1.5458807769591427, "grad_norm": 2.9646469016852617, "learning_rate": 3.8514902714255725e-06, "loss": 1.4174, "step": 2308 }, { "epoch": 1.5465505693235098, "grad_norm": 2.4350465866233186, "learning_rate": 3.850260525219699e-06, "loss": 1.2936, "step": 2309 }, { "epoch": 1.5472203616878768, "grad_norm": 2.395058117111707, "learning_rate": 3.849030317566662e-06, "loss": 1.3086, "step": 2310 }, { "epoch": 1.5478901540522438, "grad_norm": 2.2213254663294455, "learning_rate": 3.847799648886882e-06, "loss": 1.4263, "step": 2311 }, { "epoch": 1.5485599464166109, "grad_norm": 2.1023592851937902, "learning_rate": 3.8465685196009356e-06, "loss": 1.2909, "step": 2312 }, { "epoch": 1.549229738780978, "grad_norm": 4.852776068817309, "learning_rate": 3.845336930129557e-06, "loss": 1.3991, "step": 2313 }, { "epoch": 1.549899531145345, "grad_norm": 3.2956388883608745, "learning_rate": 3.8441048808936385e-06, "loss": 1.3902, "step": 2314 }, { "epoch": 1.550569323509712, "grad_norm": 2.3023974712689843, "learning_rate": 3.842872372314229e-06, "loss": 1.3587, "step": 2315 }, { "epoch": 1.551239115874079, "grad_norm": 3.388592648112413, "learning_rate": 3.841639404812534e-06, "loss": 1.3986, "step": 2316 }, { "epoch": 1.551908908238446, "grad_norm": 3.815878234551654, "learning_rate": 3.840405978809916e-06, "loss": 1.4875, "step": 2317 }, { "epoch": 1.552578700602813, "grad_norm": 2.924669690387121, "learning_rate": 3.839172094727895e-06, "loss": 1.1313, "step": 2318 }, { "epoch": 1.5532484929671801, "grad_norm": 4.4334364687652235, "learning_rate": 3.837937752988146e-06, "loss": 1.3764, "step": 2319 }, { "epoch": 1.5539182853315472, "grad_norm": 3.148839212184461, "learning_rate": 3.836702954012501e-06, "loss": 1.3619, "step": 2320 }, { "epoch": 1.5545880776959142, "grad_norm": 2.3930160586960025, "learning_rate": 3.8354676982229495e-06, "loss": 1.5202, "step": 2321 }, { "epoch": 1.5552578700602813, "grad_norm": 3.3723844129228735, "learning_rate": 3.834231986041637e-06, "loss": 1.3561, "step": 2322 }, { "epoch": 1.5559276624246483, "grad_norm": 2.3451854620706194, "learning_rate": 3.832995817890862e-06, "loss": 1.3606, "step": 2323 }, { "epoch": 1.5565974547890153, "grad_norm": 3.1182294448501677, "learning_rate": 3.831759194193082e-06, "loss": 1.1744, "step": 2324 }, { "epoch": 1.5572672471533826, "grad_norm": 2.0586019957887234, "learning_rate": 3.83052211537091e-06, "loss": 1.4168, "step": 2325 }, { "epoch": 1.5579370395177494, "grad_norm": 2.175625664612301, "learning_rate": 3.829284581847112e-06, "loss": 1.2941, "step": 2326 }, { "epoch": 1.5586068318821167, "grad_norm": 4.338545731738127, "learning_rate": 3.828046594044612e-06, "loss": 1.2073, "step": 2327 }, { "epoch": 1.5592766242464835, "grad_norm": 3.158065005150178, "learning_rate": 3.8268081523864896e-06, "loss": 1.2585, "step": 2328 }, { "epoch": 1.5599464166108508, "grad_norm": 3.5603605273210803, "learning_rate": 3.825569257295975e-06, "loss": 1.4809, "step": 2329 }, { "epoch": 1.5606162089752176, "grad_norm": 2.367062532389882, "learning_rate": 3.82432990919646e-06, "loss": 1.4608, "step": 2330 }, { "epoch": 1.5612860013395848, "grad_norm": 2.800247130400741, "learning_rate": 3.823090108511485e-06, "loss": 1.4422, "step": 2331 }, { "epoch": 1.5619557937039517, "grad_norm": 2.1986597007147837, "learning_rate": 3.82184985566475e-06, "loss": 1.3105, "step": 2332 }, { "epoch": 1.562625586068319, "grad_norm": 2.4787261820501714, "learning_rate": 3.820609151080106e-06, "loss": 1.3903, "step": 2333 }, { "epoch": 1.5632953784326857, "grad_norm": 2.2830945357309864, "learning_rate": 3.81936799518156e-06, "loss": 1.2466, "step": 2334 }, { "epoch": 1.563965170797053, "grad_norm": 2.5394266984068046, "learning_rate": 3.8181263883932735e-06, "loss": 1.195, "step": 2335 }, { "epoch": 1.5646349631614198, "grad_norm": 2.0686414545679184, "learning_rate": 3.816884331139562e-06, "loss": 1.3609, "step": 2336 }, { "epoch": 1.565304755525787, "grad_norm": 2.725931649989302, "learning_rate": 3.8156418238448935e-06, "loss": 1.406, "step": 2337 }, { "epoch": 1.565974547890154, "grad_norm": 2.501521661212298, "learning_rate": 3.814398866933891e-06, "loss": 1.336, "step": 2338 }, { "epoch": 1.5666443402545212, "grad_norm": 5.471538121607164, "learning_rate": 3.8131554608313315e-06, "loss": 1.4887, "step": 2339 }, { "epoch": 1.5673141326188882, "grad_norm": 2.357950227201533, "learning_rate": 3.811911605962145e-06, "loss": 1.327, "step": 2340 }, { "epoch": 1.5679839249832552, "grad_norm": 2.382886131717918, "learning_rate": 3.8106673027514148e-06, "loss": 1.3789, "step": 2341 }, { "epoch": 1.5686537173476223, "grad_norm": 3.9243788928940013, "learning_rate": 3.8094225516243773e-06, "loss": 1.1593, "step": 2342 }, { "epoch": 1.5693235097119893, "grad_norm": 2.681734203873263, "learning_rate": 3.8081773530064215e-06, "loss": 1.3425, "step": 2343 }, { "epoch": 1.5699933020763563, "grad_norm": 2.2462039161042626, "learning_rate": 3.8069317073230917e-06, "loss": 1.1774, "step": 2344 }, { "epoch": 1.5706630944407234, "grad_norm": 3.292250231606588, "learning_rate": 3.805685615000082e-06, "loss": 1.4092, "step": 2345 }, { "epoch": 1.5713328868050904, "grad_norm": 4.3908122736223, "learning_rate": 3.8044390764632404e-06, "loss": 1.4122, "step": 2346 }, { "epoch": 1.5720026791694575, "grad_norm": 2.2502368897075784, "learning_rate": 3.8031920921385678e-06, "loss": 1.4317, "step": 2347 }, { "epoch": 1.5726724715338245, "grad_norm": 2.1874755383735907, "learning_rate": 3.8019446624522177e-06, "loss": 1.4612, "step": 2348 }, { "epoch": 1.5733422638981915, "grad_norm": 2.2472007065767237, "learning_rate": 3.800696787830494e-06, "loss": 1.3801, "step": 2349 }, { "epoch": 1.5740120562625586, "grad_norm": 5.008834684042207, "learning_rate": 3.7994484686998547e-06, "loss": 1.3209, "step": 2350 }, { "epoch": 1.5746818486269256, "grad_norm": 2.284334580407339, "learning_rate": 3.7981997054869076e-06, "loss": 1.2832, "step": 2351 }, { "epoch": 1.5753516409912927, "grad_norm": 2.624069943400571, "learning_rate": 3.7969504986184145e-06, "loss": 1.2269, "step": 2352 }, { "epoch": 1.5760214333556597, "grad_norm": 3.6628263822286606, "learning_rate": 3.7957008485212874e-06, "loss": 1.3241, "step": 2353 }, { "epoch": 1.576691225720027, "grad_norm": 2.2920361286044275, "learning_rate": 3.79445075562259e-06, "loss": 1.2911, "step": 2354 }, { "epoch": 1.5773610180843938, "grad_norm": 2.210347864943536, "learning_rate": 3.7932002203495386e-06, "loss": 1.4651, "step": 2355 }, { "epoch": 1.578030810448761, "grad_norm": 2.4182975239207205, "learning_rate": 3.791949243129499e-06, "loss": 1.3651, "step": 2356 }, { "epoch": 1.5787006028131279, "grad_norm": 6.186424182106943, "learning_rate": 3.7906978243899885e-06, "loss": 1.219, "step": 2357 }, { "epoch": 1.5793703951774951, "grad_norm": 2.1747113516545924, "learning_rate": 3.7894459645586745e-06, "loss": 1.3154, "step": 2358 }, { "epoch": 1.580040187541862, "grad_norm": 3.6135844106975097, "learning_rate": 3.7881936640633774e-06, "loss": 1.3645, "step": 2359 }, { "epoch": 1.5807099799062292, "grad_norm": 2.4918369939306184, "learning_rate": 3.7869409233320657e-06, "loss": 1.4028, "step": 2360 }, { "epoch": 1.581379772270596, "grad_norm": 2.398021173244431, "learning_rate": 3.7856877427928616e-06, "loss": 1.2285, "step": 2361 }, { "epoch": 1.5820495646349633, "grad_norm": 2.268649556575979, "learning_rate": 3.784434122874032e-06, "loss": 1.2788, "step": 2362 }, { "epoch": 1.58271935699933, "grad_norm": 2.456067662699402, "learning_rate": 3.7831800640040015e-06, "loss": 1.2876, "step": 2363 }, { "epoch": 1.5833891493636973, "grad_norm": 2.540262253800549, "learning_rate": 3.7819255666113375e-06, "loss": 1.3719, "step": 2364 }, { "epoch": 1.5840589417280642, "grad_norm": 2.1905280851024327, "learning_rate": 3.780670631124762e-06, "loss": 1.1493, "step": 2365 }, { "epoch": 1.5847287340924314, "grad_norm": 4.198512924656532, "learning_rate": 3.7794152579731434e-06, "loss": 1.1753, "step": 2366 }, { "epoch": 1.5853985264567982, "grad_norm": 2.192628090308588, "learning_rate": 3.7781594475855044e-06, "loss": 1.2647, "step": 2367 }, { "epoch": 1.5860683188211655, "grad_norm": 2.397505294772847, "learning_rate": 3.776903200391011e-06, "loss": 1.2908, "step": 2368 }, { "epoch": 1.5867381111855325, "grad_norm": 3.0524912174685923, "learning_rate": 3.775646516818985e-06, "loss": 1.3135, "step": 2369 }, { "epoch": 1.5874079035498996, "grad_norm": 2.8410592406069526, "learning_rate": 3.7743893972988898e-06, "loss": 1.3213, "step": 2370 }, { "epoch": 1.5880776959142666, "grad_norm": 2.308620494345217, "learning_rate": 3.7731318422603458e-06, "loss": 1.4885, "step": 2371 }, { "epoch": 1.5887474882786337, "grad_norm": 3.4027904510939844, "learning_rate": 3.7718738521331153e-06, "loss": 1.3645, "step": 2372 }, { "epoch": 1.5894172806430007, "grad_norm": 2.306344998660814, "learning_rate": 3.7706154273471146e-06, "loss": 1.3192, "step": 2373 }, { "epoch": 1.5900870730073677, "grad_norm": 2.1758749605832812, "learning_rate": 3.7693565683324052e-06, "loss": 1.3877, "step": 2374 }, { "epoch": 1.5907568653717348, "grad_norm": 4.939347167688604, "learning_rate": 3.768097275519198e-06, "loss": 1.4651, "step": 2375 }, { "epoch": 1.5914266577361018, "grad_norm": 2.371063483296837, "learning_rate": 3.766837549337853e-06, "loss": 1.4467, "step": 2376 }, { "epoch": 1.5920964501004689, "grad_norm": 2.35860333027593, "learning_rate": 3.765577390218877e-06, "loss": 1.296, "step": 2377 }, { "epoch": 1.592766242464836, "grad_norm": 2.2597020037947697, "learning_rate": 3.7643167985929246e-06, "loss": 1.335, "step": 2378 }, { "epoch": 1.593436034829203, "grad_norm": 2.3696823172116566, "learning_rate": 3.7630557748908007e-06, "loss": 1.3923, "step": 2379 }, { "epoch": 1.59410582719357, "grad_norm": 2.5455834232240506, "learning_rate": 3.7617943195434546e-06, "loss": 1.2561, "step": 2380 }, { "epoch": 1.594775619557937, "grad_norm": 2.355975910830898, "learning_rate": 3.7605324329819858e-06, "loss": 1.2048, "step": 2381 }, { "epoch": 1.595445411922304, "grad_norm": 2.176689102597227, "learning_rate": 3.7592701156376378e-06, "loss": 1.2418, "step": 2382 }, { "epoch": 1.596115204286671, "grad_norm": 3.1783396757807596, "learning_rate": 3.758007367941806e-06, "loss": 1.1067, "step": 2383 }, { "epoch": 1.5967849966510381, "grad_norm": 2.559638636406259, "learning_rate": 3.756744190326028e-06, "loss": 1.3973, "step": 2384 }, { "epoch": 1.5974547890154054, "grad_norm": 2.1785346566284667, "learning_rate": 3.755480583221993e-06, "loss": 1.3295, "step": 2385 }, { "epoch": 1.5981245813797722, "grad_norm": 2.227673774672654, "learning_rate": 3.754216547061533e-06, "loss": 1.32, "step": 2386 }, { "epoch": 1.5987943737441395, "grad_norm": 2.319832626103924, "learning_rate": 3.7529520822766295e-06, "loss": 1.3277, "step": 2387 }, { "epoch": 1.5994641661085063, "grad_norm": 4.256987331073695, "learning_rate": 3.751687189299408e-06, "loss": 1.3362, "step": 2388 }, { "epoch": 1.6001339584728735, "grad_norm": 2.5210386320914058, "learning_rate": 3.7504218685621417e-06, "loss": 1.4145, "step": 2389 }, { "epoch": 1.6008037508372404, "grad_norm": 2.912169023476226, "learning_rate": 3.7491561204972504e-06, "loss": 1.4246, "step": 2390 }, { "epoch": 1.6014735432016076, "grad_norm": 2.5018288238338795, "learning_rate": 3.7478899455372997e-06, "loss": 1.3882, "step": 2391 }, { "epoch": 1.6021433355659744, "grad_norm": 2.226611029261743, "learning_rate": 3.746623344115e-06, "loss": 1.4468, "step": 2392 }, { "epoch": 1.6028131279303417, "grad_norm": 2.207985363184561, "learning_rate": 3.745356316663209e-06, "loss": 1.2266, "step": 2393 }, { "epoch": 1.6034829202947085, "grad_norm": 2.991745967052074, "learning_rate": 3.744088863614928e-06, "loss": 1.4342, "step": 2394 }, { "epoch": 1.6041527126590758, "grad_norm": 2.4085293738754405, "learning_rate": 3.7428209854033065e-06, "loss": 1.423, "step": 2395 }, { "epoch": 1.6048225050234426, "grad_norm": 2.1987505138497836, "learning_rate": 3.7415526824616367e-06, "loss": 1.3304, "step": 2396 }, { "epoch": 1.6054922973878099, "grad_norm": 3.5035566741286077, "learning_rate": 3.740283955223357e-06, "loss": 1.1788, "step": 2397 }, { "epoch": 1.6061620897521767, "grad_norm": 2.355269333463272, "learning_rate": 3.739014804122052e-06, "loss": 1.3744, "step": 2398 }, { "epoch": 1.606831882116544, "grad_norm": 4.1940161742809074, "learning_rate": 3.7377452295914487e-06, "loss": 1.1874, "step": 2399 }, { "epoch": 1.607501674480911, "grad_norm": 2.863319959182034, "learning_rate": 3.736475232065421e-06, "loss": 1.2374, "step": 2400 }, { "epoch": 1.608171466845278, "grad_norm": 2.8913389139894576, "learning_rate": 3.7352048119779855e-06, "loss": 1.3603, "step": 2401 }, { "epoch": 1.608841259209645, "grad_norm": 2.2092359923097065, "learning_rate": 3.7339339697633053e-06, "loss": 1.3543, "step": 2402 }, { "epoch": 1.609511051574012, "grad_norm": 3.1736671198800432, "learning_rate": 3.7326627058556854e-06, "loss": 1.3124, "step": 2403 }, { "epoch": 1.6101808439383791, "grad_norm": 6.064527584998624, "learning_rate": 3.7313910206895776e-06, "loss": 1.5131, "step": 2404 }, { "epoch": 1.6108506363027462, "grad_norm": 2.164541838394903, "learning_rate": 3.7301189146995737e-06, "loss": 1.2227, "step": 2405 }, { "epoch": 1.6115204286671132, "grad_norm": 2.1944289516435367, "learning_rate": 3.728846388320415e-06, "loss": 1.2127, "step": 2406 }, { "epoch": 1.6121902210314802, "grad_norm": 3.507406884368241, "learning_rate": 3.7275734419869804e-06, "loss": 1.3694, "step": 2407 }, { "epoch": 1.6128600133958473, "grad_norm": 3.2674640513027993, "learning_rate": 3.7263000761342976e-06, "loss": 1.0796, "step": 2408 }, { "epoch": 1.6135298057602143, "grad_norm": 2.7795546581578083, "learning_rate": 3.7250262911975333e-06, "loss": 1.2726, "step": 2409 }, { "epoch": 1.6141995981245814, "grad_norm": 3.1292653700100446, "learning_rate": 3.723752087612001e-06, "loss": 1.2003, "step": 2410 }, { "epoch": 1.6148693904889484, "grad_norm": 2.326041111516032, "learning_rate": 3.7224774658131536e-06, "loss": 1.3179, "step": 2411 }, { "epoch": 1.6155391828533154, "grad_norm": 2.1860221770149435, "learning_rate": 3.7212024262365913e-06, "loss": 1.4492, "step": 2412 }, { "epoch": 1.6162089752176825, "grad_norm": 2.3781604976575887, "learning_rate": 3.719926969318053e-06, "loss": 1.4104, "step": 2413 }, { "epoch": 1.6168787675820495, "grad_norm": 2.8337510604687717, "learning_rate": 3.7186510954934236e-06, "loss": 1.2327, "step": 2414 }, { "epoch": 1.6175485599464166, "grad_norm": 2.280269498921208, "learning_rate": 3.7173748051987264e-06, "loss": 1.3187, "step": 2415 }, { "epoch": 1.6182183523107838, "grad_norm": 2.7802353376240987, "learning_rate": 3.7160980988701323e-06, "loss": 1.386, "step": 2416 }, { "epoch": 1.6188881446751506, "grad_norm": 3.329088319024179, "learning_rate": 3.714820976943949e-06, "loss": 1.1949, "step": 2417 }, { "epoch": 1.619557937039518, "grad_norm": 2.2205554075182357, "learning_rate": 3.7135434398566318e-06, "loss": 1.3836, "step": 2418 }, { "epoch": 1.6202277294038847, "grad_norm": 3.172444542306734, "learning_rate": 3.712265488044771e-06, "loss": 1.2437, "step": 2419 }, { "epoch": 1.620897521768252, "grad_norm": 2.3456774437989845, "learning_rate": 3.710987121945106e-06, "loss": 1.4705, "step": 2420 }, { "epoch": 1.6215673141326188, "grad_norm": 2.7113469905141576, "learning_rate": 3.709708341994512e-06, "loss": 1.3831, "step": 2421 }, { "epoch": 1.622237106496986, "grad_norm": 2.3360897367691678, "learning_rate": 3.7084291486300093e-06, "loss": 1.239, "step": 2422 }, { "epoch": 1.6229068988613529, "grad_norm": 2.0961026278235853, "learning_rate": 3.7071495422887573e-06, "loss": 1.3359, "step": 2423 }, { "epoch": 1.6235766912257201, "grad_norm": 2.2974697717534074, "learning_rate": 3.7058695234080577e-06, "loss": 1.2517, "step": 2424 }, { "epoch": 1.624246483590087, "grad_norm": 2.160997872363273, "learning_rate": 3.704589092425352e-06, "loss": 1.3892, "step": 2425 }, { "epoch": 1.6249162759544542, "grad_norm": 3.0322465132204375, "learning_rate": 3.7033082497782246e-06, "loss": 1.1835, "step": 2426 }, { "epoch": 1.625586068318821, "grad_norm": 2.4107231227379997, "learning_rate": 3.7020269959043988e-06, "loss": 1.2664, "step": 2427 }, { "epoch": 1.6262558606831883, "grad_norm": 2.5849936520799126, "learning_rate": 3.7007453312417383e-06, "loss": 1.4196, "step": 2428 }, { "epoch": 1.626925653047555, "grad_norm": 2.2979462865637235, "learning_rate": 3.699463256228249e-06, "loss": 1.3451, "step": 2429 }, { "epoch": 1.6275954454119224, "grad_norm": 2.0509696253608696, "learning_rate": 3.698180771302075e-06, "loss": 1.385, "step": 2430 }, { "epoch": 1.6282652377762894, "grad_norm": 2.5850873540278685, "learning_rate": 3.696897876901502e-06, "loss": 1.4789, "step": 2431 }, { "epoch": 1.6289350301406564, "grad_norm": 2.33658443231377, "learning_rate": 3.6956145734649547e-06, "loss": 1.2067, "step": 2432 }, { "epoch": 1.6296048225050235, "grad_norm": 2.6690929347503145, "learning_rate": 3.694330861430998e-06, "loss": 1.5167, "step": 2433 }, { "epoch": 1.6302746148693905, "grad_norm": 2.1913047783564643, "learning_rate": 3.6930467412383357e-06, "loss": 1.215, "step": 2434 }, { "epoch": 1.6309444072337576, "grad_norm": 2.4174457688995608, "learning_rate": 3.6917622133258125e-06, "loss": 1.4872, "step": 2435 }, { "epoch": 1.6316141995981246, "grad_norm": 2.319035407799145, "learning_rate": 3.6904772781324116e-06, "loss": 1.3385, "step": 2436 }, { "epoch": 1.6322839919624916, "grad_norm": 3.544543582645246, "learning_rate": 3.689191936097255e-06, "loss": 1.2391, "step": 2437 }, { "epoch": 1.6329537843268587, "grad_norm": 3.213721430684574, "learning_rate": 3.6879061876596044e-06, "loss": 1.346, "step": 2438 }, { "epoch": 1.6336235766912257, "grad_norm": 2.3307882821432995, "learning_rate": 3.686620033258861e-06, "loss": 1.3199, "step": 2439 }, { "epoch": 1.6342933690555927, "grad_norm": 2.2164323944623354, "learning_rate": 3.6853334733345626e-06, "loss": 1.3476, "step": 2440 }, { "epoch": 1.6349631614199598, "grad_norm": 2.4632333376850455, "learning_rate": 3.684046508326387e-06, "loss": 1.3284, "step": 2441 }, { "epoch": 1.6356329537843268, "grad_norm": 2.170308572534225, "learning_rate": 3.68275913867415e-06, "loss": 1.2047, "step": 2442 }, { "epoch": 1.6363027461486939, "grad_norm": 2.565762331552281, "learning_rate": 3.6814713648178087e-06, "loss": 1.2022, "step": 2443 }, { "epoch": 1.636972538513061, "grad_norm": 2.9873244867004956, "learning_rate": 3.680183187197452e-06, "loss": 1.2024, "step": 2444 }, { "epoch": 1.637642330877428, "grad_norm": 6.076635920154763, "learning_rate": 3.678894606253313e-06, "loss": 1.2999, "step": 2445 }, { "epoch": 1.638312123241795, "grad_norm": 2.475322502417628, "learning_rate": 3.677605622425759e-06, "loss": 1.2457, "step": 2446 }, { "epoch": 1.6389819156061622, "grad_norm": 2.1517781641098166, "learning_rate": 3.676316236155296e-06, "loss": 1.2276, "step": 2447 }, { "epoch": 1.639651707970529, "grad_norm": 2.5104297567680494, "learning_rate": 3.6750264478825675e-06, "loss": 1.2433, "step": 2448 }, { "epoch": 1.6403215003348963, "grad_norm": 3.747839953472579, "learning_rate": 3.673736258048356e-06, "loss": 1.4205, "step": 2449 }, { "epoch": 1.6409912926992631, "grad_norm": 2.486116533904712, "learning_rate": 3.672445667093577e-06, "loss": 1.3127, "step": 2450 }, { "epoch": 1.6416610850636304, "grad_norm": 5.094598518433392, "learning_rate": 3.6711546754592885e-06, "loss": 1.2887, "step": 2451 }, { "epoch": 1.6423308774279972, "grad_norm": 2.700553901519732, "learning_rate": 3.6698632835866805e-06, "loss": 1.5176, "step": 2452 }, { "epoch": 1.6430006697923645, "grad_norm": 2.3864919230284594, "learning_rate": 3.6685714919170834e-06, "loss": 1.4234, "step": 2453 }, { "epoch": 1.6436704621567313, "grad_norm": 2.7749593737745473, "learning_rate": 3.667279300891962e-06, "loss": 1.1826, "step": 2454 }, { "epoch": 1.6443402545210986, "grad_norm": 5.886989248650033, "learning_rate": 3.6659867109529197e-06, "loss": 1.3546, "step": 2455 }, { "epoch": 1.6450100468854654, "grad_norm": 3.2695293100373526, "learning_rate": 3.664693722541692e-06, "loss": 1.0626, "step": 2456 }, { "epoch": 1.6456798392498326, "grad_norm": 3.013122082735927, "learning_rate": 3.663400336100158e-06, "loss": 1.3438, "step": 2457 }, { "epoch": 1.6463496316141994, "grad_norm": 2.274891165078473, "learning_rate": 3.662106552070324e-06, "loss": 1.2564, "step": 2458 }, { "epoch": 1.6470194239785667, "grad_norm": 2.7405944175811623, "learning_rate": 3.6608123708943393e-06, "loss": 1.3749, "step": 2459 }, { "epoch": 1.6476892163429335, "grad_norm": 3.4423424297690093, "learning_rate": 3.659517793014484e-06, "loss": 1.3065, "step": 2460 }, { "epoch": 1.6483590087073008, "grad_norm": 2.1689907280001193, "learning_rate": 3.6582228188731783e-06, "loss": 1.3785, "step": 2461 }, { "epoch": 1.6490288010716678, "grad_norm": 2.2285858055862744, "learning_rate": 3.6569274489129737e-06, "loss": 1.2414, "step": 2462 }, { "epoch": 1.6496985934360349, "grad_norm": 4.213634930915533, "learning_rate": 3.6556316835765593e-06, "loss": 1.192, "step": 2463 }, { "epoch": 1.650368385800402, "grad_norm": 2.788864734533914, "learning_rate": 3.6543355233067586e-06, "loss": 1.3931, "step": 2464 }, { "epoch": 1.651038178164769, "grad_norm": 2.481322731732739, "learning_rate": 3.6530389685465297e-06, "loss": 1.158, "step": 2465 }, { "epoch": 1.651707970529136, "grad_norm": 2.217217245311968, "learning_rate": 3.651742019738967e-06, "loss": 1.3134, "step": 2466 }, { "epoch": 1.652377762893503, "grad_norm": 2.229876865011373, "learning_rate": 3.6504446773272978e-06, "loss": 1.4342, "step": 2467 }, { "epoch": 1.65304755525787, "grad_norm": 2.9121697803742506, "learning_rate": 3.6491469417548845e-06, "loss": 1.3314, "step": 2468 }, { "epoch": 1.653717347622237, "grad_norm": 2.821812938613957, "learning_rate": 3.6478488134652248e-06, "loss": 1.2643, "step": 2469 }, { "epoch": 1.6543871399866041, "grad_norm": 4.626449218884552, "learning_rate": 3.646550292901949e-06, "loss": 1.3712, "step": 2470 }, { "epoch": 1.6550569323509712, "grad_norm": 3.2145280359425774, "learning_rate": 3.6452513805088236e-06, "loss": 1.2106, "step": 2471 }, { "epoch": 1.6557267247153382, "grad_norm": 2.4126061647256916, "learning_rate": 3.643952076729746e-06, "loss": 1.4117, "step": 2472 }, { "epoch": 1.6563965170797053, "grad_norm": 2.557518002280652, "learning_rate": 3.64265238200875e-06, "loss": 1.1604, "step": 2473 }, { "epoch": 1.6570663094440723, "grad_norm": 2.192140313136034, "learning_rate": 3.641352296790003e-06, "loss": 1.4108, "step": 2474 }, { "epoch": 1.6577361018084393, "grad_norm": 2.263825190794258, "learning_rate": 3.640051821517803e-06, "loss": 1.3316, "step": 2475 }, { "epoch": 1.6584058941728066, "grad_norm": 2.0583674603477427, "learning_rate": 3.6387509566365836e-06, "loss": 1.3749, "step": 2476 }, { "epoch": 1.6590756865371734, "grad_norm": 2.016382818243122, "learning_rate": 3.6374497025909124e-06, "loss": 1.1746, "step": 2477 }, { "epoch": 1.6597454789015407, "grad_norm": 2.064629190579179, "learning_rate": 3.636148059825487e-06, "loss": 1.2169, "step": 2478 }, { "epoch": 1.6604152712659075, "grad_norm": 2.0301721791908913, "learning_rate": 3.6348460287851415e-06, "loss": 1.3524, "step": 2479 }, { "epoch": 1.6610850636302747, "grad_norm": 3.9309819622559834, "learning_rate": 3.63354360991484e-06, "loss": 1.3727, "step": 2480 }, { "epoch": 1.6617548559946416, "grad_norm": 2.3025774308374856, "learning_rate": 3.6322408036596803e-06, "loss": 1.3578, "step": 2481 }, { "epoch": 1.6624246483590088, "grad_norm": 2.3426212437470184, "learning_rate": 3.6309376104648907e-06, "loss": 1.2706, "step": 2482 }, { "epoch": 1.6630944407233756, "grad_norm": 2.387616233805459, "learning_rate": 3.6296340307758347e-06, "loss": 1.4471, "step": 2483 }, { "epoch": 1.663764233087743, "grad_norm": 4.2642801509758055, "learning_rate": 3.628330065038006e-06, "loss": 1.1726, "step": 2484 }, { "epoch": 1.6644340254521097, "grad_norm": 2.369385505750417, "learning_rate": 3.6270257136970304e-06, "loss": 1.3994, "step": 2485 }, { "epoch": 1.665103817816477, "grad_norm": 2.9750199815367298, "learning_rate": 3.625720977198667e-06, "loss": 1.3709, "step": 2486 }, { "epoch": 1.6657736101808438, "grad_norm": 2.3346658816720662, "learning_rate": 3.6244158559888025e-06, "loss": 1.269, "step": 2487 }, { "epoch": 1.666443402545211, "grad_norm": 2.0459338795583704, "learning_rate": 3.6231103505134618e-06, "loss": 1.0719, "step": 2488 }, { "epoch": 1.6671131949095779, "grad_norm": 2.3888640263822816, "learning_rate": 3.6218044612187935e-06, "loss": 1.3822, "step": 2489 }, { "epoch": 1.6677829872739451, "grad_norm": 2.2440424648949295, "learning_rate": 3.6204981885510835e-06, "loss": 1.2858, "step": 2490 }, { "epoch": 1.6684527796383122, "grad_norm": 2.9146553179388293, "learning_rate": 3.619191532956745e-06, "loss": 1.3383, "step": 2491 }, { "epoch": 1.6691225720026792, "grad_norm": 3.3434441676256914, "learning_rate": 3.6178844948823246e-06, "loss": 1.3662, "step": 2492 }, { "epoch": 1.6697923643670463, "grad_norm": 2.127903660460959, "learning_rate": 3.6165770747744955e-06, "loss": 1.2178, "step": 2493 }, { "epoch": 1.6704621567314133, "grad_norm": 3.056066203954961, "learning_rate": 3.6152692730800677e-06, "loss": 1.1834, "step": 2494 }, { "epoch": 1.6711319490957803, "grad_norm": 3.3984732609830477, "learning_rate": 3.613961090245976e-06, "loss": 1.3499, "step": 2495 }, { "epoch": 1.6718017414601474, "grad_norm": 4.348881462550222, "learning_rate": 3.612652526719289e-06, "loss": 1.2906, "step": 2496 }, { "epoch": 1.6724715338245144, "grad_norm": 2.5595743889104012, "learning_rate": 3.611343582947202e-06, "loss": 1.4734, "step": 2497 }, { "epoch": 1.6731413261888815, "grad_norm": 2.3944849716038847, "learning_rate": 3.610034259377045e-06, "loss": 1.4312, "step": 2498 }, { "epoch": 1.6738111185532485, "grad_norm": 2.551329458620574, "learning_rate": 3.6087245564562733e-06, "loss": 1.3736, "step": 2499 }, { "epoch": 1.6744809109176155, "grad_norm": 3.550385819282981, "learning_rate": 3.607414474632474e-06, "loss": 1.398, "step": 2500 }, { "epoch": 1.6751507032819826, "grad_norm": 2.4519646110372237, "learning_rate": 3.606104014353364e-06, "loss": 1.3768, "step": 2501 }, { "epoch": 1.6758204956463496, "grad_norm": 2.941230605450344, "learning_rate": 3.6047931760667877e-06, "loss": 1.2291, "step": 2502 }, { "epoch": 1.6764902880107166, "grad_norm": 2.8117048160117126, "learning_rate": 3.6034819602207204e-06, "loss": 1.1189, "step": 2503 }, { "epoch": 1.6771600803750837, "grad_norm": 2.20515680726053, "learning_rate": 3.6021703672632657e-06, "loss": 1.387, "step": 2504 }, { "epoch": 1.6778298727394507, "grad_norm": 2.3574496148823014, "learning_rate": 3.600858397642657e-06, "loss": 1.2985, "step": 2505 }, { "epoch": 1.6784996651038178, "grad_norm": 3.209180869148051, "learning_rate": 3.5995460518072545e-06, "loss": 1.2037, "step": 2506 }, { "epoch": 1.679169457468185, "grad_norm": 3.1349072778103793, "learning_rate": 3.5982333302055496e-06, "loss": 1.3477, "step": 2507 }, { "epoch": 1.6798392498325518, "grad_norm": 2.7370594205061702, "learning_rate": 3.5969202332861593e-06, "loss": 1.2381, "step": 2508 }, { "epoch": 1.680509042196919, "grad_norm": 2.477024254514927, "learning_rate": 3.5956067614978307e-06, "loss": 1.2018, "step": 2509 }, { "epoch": 1.681178834561286, "grad_norm": 2.8291741975730877, "learning_rate": 3.5942929152894386e-06, "loss": 1.1982, "step": 2510 }, { "epoch": 1.6818486269256532, "grad_norm": 2.257023068121091, "learning_rate": 3.5929786951099856e-06, "loss": 1.243, "step": 2511 }, { "epoch": 1.68251841929002, "grad_norm": 2.2513289939195156, "learning_rate": 3.591664101408602e-06, "loss": 1.3186, "step": 2512 }, { "epoch": 1.6831882116543873, "grad_norm": 2.559887257948032, "learning_rate": 3.5903491346345466e-06, "loss": 1.3226, "step": 2513 }, { "epoch": 1.683858004018754, "grad_norm": 4.123263961134172, "learning_rate": 3.5890337952372045e-06, "loss": 1.3514, "step": 2514 }, { "epoch": 1.6845277963831213, "grad_norm": 2.076576423712208, "learning_rate": 3.5877180836660885e-06, "loss": 1.2803, "step": 2515 }, { "epoch": 1.6851975887474882, "grad_norm": 2.4358006186751164, "learning_rate": 3.5864020003708392e-06, "loss": 1.3682, "step": 2516 }, { "epoch": 1.6858673811118554, "grad_norm": 2.2874825693040868, "learning_rate": 3.585085545801224e-06, "loss": 1.367, "step": 2517 }, { "epoch": 1.6865371734762222, "grad_norm": 4.248156568972568, "learning_rate": 3.5837687204071365e-06, "loss": 1.4667, "step": 2518 }, { "epoch": 1.6872069658405895, "grad_norm": 2.3877873397108873, "learning_rate": 3.582451524638597e-06, "loss": 1.4126, "step": 2519 }, { "epoch": 1.6878767582049563, "grad_norm": 3.6404858034206398, "learning_rate": 3.5811339589457543e-06, "loss": 1.4018, "step": 2520 }, { "epoch": 1.6885465505693236, "grad_norm": 2.1886886347047794, "learning_rate": 3.5798160237788813e-06, "loss": 1.3855, "step": 2521 }, { "epoch": 1.6892163429336906, "grad_norm": 2.1381626906546005, "learning_rate": 3.5784977195883787e-06, "loss": 1.2969, "step": 2522 }, { "epoch": 1.6898861352980576, "grad_norm": 2.1615949782479777, "learning_rate": 3.5771790468247715e-06, "loss": 1.2354, "step": 2523 }, { "epoch": 1.6905559276624247, "grad_norm": 3.8143335984909967, "learning_rate": 3.575860005938713e-06, "loss": 1.2659, "step": 2524 }, { "epoch": 1.6912257200267917, "grad_norm": 6.24886582621189, "learning_rate": 3.5745405973809806e-06, "loss": 1.1115, "step": 2525 }, { "epoch": 1.6918955123911588, "grad_norm": 2.4634158369993524, "learning_rate": 3.5732208216024776e-06, "loss": 1.3172, "step": 2526 }, { "epoch": 1.6925653047555258, "grad_norm": 3.643334257444332, "learning_rate": 3.5719006790542343e-06, "loss": 1.3031, "step": 2527 }, { "epoch": 1.6932350971198928, "grad_norm": 2.4344248714516485, "learning_rate": 3.570580170187404e-06, "loss": 1.4175, "step": 2528 }, { "epoch": 1.6939048894842599, "grad_norm": 2.9806167026137866, "learning_rate": 3.569259295453267e-06, "loss": 1.2607, "step": 2529 }, { "epoch": 1.694574681848627, "grad_norm": 5.113711240584224, "learning_rate": 3.567938055303227e-06, "loss": 1.2301, "step": 2530 }, { "epoch": 1.695244474212994, "grad_norm": 3.8016864575726275, "learning_rate": 3.5666164501888147e-06, "loss": 1.3541, "step": 2531 }, { "epoch": 1.695914266577361, "grad_norm": 2.168734067256422, "learning_rate": 3.565294480561683e-06, "loss": 1.2946, "step": 2532 }, { "epoch": 1.696584058941728, "grad_norm": 2.4175374958716613, "learning_rate": 3.563972146873613e-06, "loss": 1.3636, "step": 2533 }, { "epoch": 1.697253851306095, "grad_norm": 3.6457136840240603, "learning_rate": 3.562649449576505e-06, "loss": 1.2512, "step": 2534 }, { "epoch": 1.6979236436704621, "grad_norm": 2.859223236196285, "learning_rate": 3.561326389122389e-06, "loss": 1.2137, "step": 2535 }, { "epoch": 1.6985934360348292, "grad_norm": 2.6881172666635575, "learning_rate": 3.560002965963416e-06, "loss": 1.2352, "step": 2536 }, { "epoch": 1.6992632283991962, "grad_norm": 2.5793724277744094, "learning_rate": 3.5586791805518616e-06, "loss": 1.13, "step": 2537 }, { "epoch": 1.6999330207635635, "grad_norm": 2.3926971553604086, "learning_rate": 3.5573550333401242e-06, "loss": 1.3665, "step": 2538 }, { "epoch": 1.7006028131279303, "grad_norm": 2.748447490765736, "learning_rate": 3.5560305247807273e-06, "loss": 1.4133, "step": 2539 }, { "epoch": 1.7012726054922975, "grad_norm": 4.002359831056547, "learning_rate": 3.554705655326318e-06, "loss": 1.261, "step": 2540 }, { "epoch": 1.7019423978566643, "grad_norm": 2.3948978137815935, "learning_rate": 3.553380425429666e-06, "loss": 1.2918, "step": 2541 }, { "epoch": 1.7026121902210316, "grad_norm": 3.0357395537864993, "learning_rate": 3.552054835543664e-06, "loss": 1.1495, "step": 2542 }, { "epoch": 1.7032819825853984, "grad_norm": 5.3245526319836225, "learning_rate": 3.550728886121328e-06, "loss": 1.2617, "step": 2543 }, { "epoch": 1.7039517749497657, "grad_norm": 2.2752753135674735, "learning_rate": 3.549402577615797e-06, "loss": 1.3318, "step": 2544 }, { "epoch": 1.7046215673141325, "grad_norm": 3.3982508366740523, "learning_rate": 3.548075910480333e-06, "loss": 1.1816, "step": 2545 }, { "epoch": 1.7052913596784998, "grad_norm": 2.3828508254135574, "learning_rate": 3.5467488851683195e-06, "loss": 1.3822, "step": 2546 }, { "epoch": 1.7059611520428666, "grad_norm": 2.338030961151524, "learning_rate": 3.545421502133264e-06, "loss": 1.1983, "step": 2547 }, { "epoch": 1.7066309444072338, "grad_norm": 2.640879662374246, "learning_rate": 3.544093761828794e-06, "loss": 1.2231, "step": 2548 }, { "epoch": 1.7073007367716007, "grad_norm": 3.042620066466561, "learning_rate": 3.542765664708662e-06, "loss": 1.3092, "step": 2549 }, { "epoch": 1.707970529135968, "grad_norm": 2.3137619670247593, "learning_rate": 3.54143721122674e-06, "loss": 1.2669, "step": 2550 }, { "epoch": 1.7086403215003347, "grad_norm": 3.7282055612983185, "learning_rate": 3.540108401837023e-06, "loss": 1.1684, "step": 2551 }, { "epoch": 1.709310113864702, "grad_norm": 2.138432123070715, "learning_rate": 3.5387792369936268e-06, "loss": 1.2759, "step": 2552 }, { "epoch": 1.709979906229069, "grad_norm": 2.415172855711182, "learning_rate": 3.5374497171507897e-06, "loss": 1.4649, "step": 2553 }, { "epoch": 1.710649698593436, "grad_norm": 2.215890242528263, "learning_rate": 3.5361198427628706e-06, "loss": 1.2164, "step": 2554 }, { "epoch": 1.7113194909578031, "grad_norm": 2.2072337149295005, "learning_rate": 3.534789614284351e-06, "loss": 1.3385, "step": 2555 }, { "epoch": 1.7119892833221702, "grad_norm": 2.6870190558746287, "learning_rate": 3.533459032169829e-06, "loss": 1.3318, "step": 2556 }, { "epoch": 1.7126590756865372, "grad_norm": 2.185418176230234, "learning_rate": 3.5321280968740303e-06, "loss": 1.3611, "step": 2557 }, { "epoch": 1.7133288680509042, "grad_norm": 2.119155659394187, "learning_rate": 3.5307968088517954e-06, "loss": 1.4326, "step": 2558 }, { "epoch": 1.7139986604152713, "grad_norm": 2.318662790269956, "learning_rate": 3.529465168558089e-06, "loss": 1.2165, "step": 2559 }, { "epoch": 1.7146684527796383, "grad_norm": 2.0840453947459303, "learning_rate": 3.528133176447994e-06, "loss": 1.3105, "step": 2560 }, { "epoch": 1.7153382451440053, "grad_norm": 3.34486491476412, "learning_rate": 3.5268008329767144e-06, "loss": 1.3298, "step": 2561 }, { "epoch": 1.7160080375083724, "grad_norm": 2.4804092420622754, "learning_rate": 3.525468138599575e-06, "loss": 1.0818, "step": 2562 }, { "epoch": 1.7166778298727394, "grad_norm": 2.757450167994697, "learning_rate": 3.524135093772019e-06, "loss": 1.2796, "step": 2563 }, { "epoch": 1.7173476222371065, "grad_norm": 3.0862535851866206, "learning_rate": 3.5228016989496107e-06, "loss": 1.2484, "step": 2564 }, { "epoch": 1.7180174146014735, "grad_norm": 2.1761285911389914, "learning_rate": 3.5214679545880332e-06, "loss": 1.2349, "step": 2565 }, { "epoch": 1.7186872069658405, "grad_norm": 3.0180532623685905, "learning_rate": 3.520133861143089e-06, "loss": 1.3672, "step": 2566 }, { "epoch": 1.7193569993302076, "grad_norm": 2.1676532845062506, "learning_rate": 3.5187994190707016e-06, "loss": 1.2837, "step": 2567 }, { "epoch": 1.7200267916945746, "grad_norm": 2.30456570507901, "learning_rate": 3.5174646288269106e-06, "loss": 1.1554, "step": 2568 }, { "epoch": 1.7206965840589419, "grad_norm": 2.3846321954347176, "learning_rate": 3.516129490867877e-06, "loss": 1.2455, "step": 2569 }, { "epoch": 1.7213663764233087, "grad_norm": 2.271955460532023, "learning_rate": 3.5147940056498788e-06, "loss": 1.3238, "step": 2570 }, { "epoch": 1.722036168787676, "grad_norm": 2.555028254790511, "learning_rate": 3.513458173629316e-06, "loss": 1.2672, "step": 2571 }, { "epoch": 1.7227059611520428, "grad_norm": 2.229411843550307, "learning_rate": 3.5121219952627027e-06, "loss": 1.0688, "step": 2572 }, { "epoch": 1.72337575351641, "grad_norm": 4.162804599973216, "learning_rate": 3.5107854710066757e-06, "loss": 1.1065, "step": 2573 }, { "epoch": 1.7240455458807769, "grad_norm": 2.3754019461188616, "learning_rate": 3.5094486013179853e-06, "loss": 1.3053, "step": 2574 }, { "epoch": 1.7247153382451441, "grad_norm": 6.174106664466813, "learning_rate": 3.5081113866535045e-06, "loss": 1.3588, "step": 2575 }, { "epoch": 1.725385130609511, "grad_norm": 4.758248260650211, "learning_rate": 3.5067738274702202e-06, "loss": 1.3667, "step": 2576 }, { "epoch": 1.7260549229738782, "grad_norm": 2.5542759514795144, "learning_rate": 3.5054359242252407e-06, "loss": 1.2509, "step": 2577 }, { "epoch": 1.726724715338245, "grad_norm": 2.1753379590852693, "learning_rate": 3.504097677375789e-06, "loss": 1.3281, "step": 2578 }, { "epoch": 1.7273945077026123, "grad_norm": 3.142639936688164, "learning_rate": 3.5027590873792074e-06, "loss": 1.2892, "step": 2579 }, { "epoch": 1.728064300066979, "grad_norm": 1.9905651654366154, "learning_rate": 3.501420154692954e-06, "loss": 1.3182, "step": 2580 }, { "epoch": 1.7287340924313463, "grad_norm": 2.1903677073947283, "learning_rate": 3.500080879774605e-06, "loss": 1.1841, "step": 2581 }, { "epoch": 1.7294038847957132, "grad_norm": 2.363824758678384, "learning_rate": 3.4987412630818534e-06, "loss": 1.3609, "step": 2582 }, { "epoch": 1.7300736771600804, "grad_norm": 3.0149252342634703, "learning_rate": 3.4974013050725087e-06, "loss": 1.3272, "step": 2583 }, { "epoch": 1.7307434695244475, "grad_norm": 2.164438094059879, "learning_rate": 3.496061006204497e-06, "loss": 1.2795, "step": 2584 }, { "epoch": 1.7314132618888145, "grad_norm": 2.1722048639868796, "learning_rate": 3.4947203669358615e-06, "loss": 1.4299, "step": 2585 }, { "epoch": 1.7320830542531815, "grad_norm": 3.0650124809515775, "learning_rate": 3.4933793877247614e-06, "loss": 1.2606, "step": 2586 }, { "epoch": 1.7327528466175486, "grad_norm": 2.764130844381554, "learning_rate": 3.4920380690294714e-06, "loss": 1.4269, "step": 2587 }, { "epoch": 1.7334226389819156, "grad_norm": 1.9481413033536323, "learning_rate": 3.490696411308383e-06, "loss": 1.2674, "step": 2588 }, { "epoch": 1.7340924313462827, "grad_norm": 2.1732258302956278, "learning_rate": 3.4893544150200043e-06, "loss": 1.2366, "step": 2589 }, { "epoch": 1.7347622237106497, "grad_norm": 2.0491135564813496, "learning_rate": 3.488012080622957e-06, "loss": 1.086, "step": 2590 }, { "epoch": 1.7354320160750167, "grad_norm": 3.0811825455807242, "learning_rate": 3.48666940857598e-06, "loss": 1.404, "step": 2591 }, { "epoch": 1.7361018084393838, "grad_norm": 2.9086879863546167, "learning_rate": 3.4853263993379272e-06, "loss": 1.2822, "step": 2592 }, { "epoch": 1.7367716008037508, "grad_norm": 2.365701116882373, "learning_rate": 3.483983053367767e-06, "loss": 1.1979, "step": 2593 }, { "epoch": 1.7374413931681179, "grad_norm": 2.5339558303434195, "learning_rate": 3.4826393711245847e-06, "loss": 1.3965, "step": 2594 }, { "epoch": 1.738111185532485, "grad_norm": 4.402093778145251, "learning_rate": 3.4812953530675784e-06, "loss": 1.2957, "step": 2595 }, { "epoch": 1.738780977896852, "grad_norm": 2.053996625873645, "learning_rate": 3.4799509996560625e-06, "loss": 1.2685, "step": 2596 }, { "epoch": 1.739450770261219, "grad_norm": 2.1174552452082605, "learning_rate": 3.4786063113494656e-06, "loss": 1.3422, "step": 2597 }, { "epoch": 1.7401205626255862, "grad_norm": 4.484498065071086, "learning_rate": 3.4772612886073297e-06, "loss": 1.3008, "step": 2598 }, { "epoch": 1.740790354989953, "grad_norm": 3.050982178854675, "learning_rate": 3.4759159318893125e-06, "loss": 1.2677, "step": 2599 }, { "epoch": 1.7414601473543203, "grad_norm": 3.7269289300673223, "learning_rate": 3.474570241655185e-06, "loss": 1.2462, "step": 2600 }, { "epoch": 1.7421299397186871, "grad_norm": 2.12073802409133, "learning_rate": 3.4732242183648336e-06, "loss": 1.3569, "step": 2601 }, { "epoch": 1.7427997320830544, "grad_norm": 3.259460592413602, "learning_rate": 3.4718778624782557e-06, "loss": 1.3089, "step": 2602 }, { "epoch": 1.7434695244474212, "grad_norm": 2.7741186652291394, "learning_rate": 3.4705311744555655e-06, "loss": 1.2026, "step": 2603 }, { "epoch": 1.7441393168117885, "grad_norm": 3.541683016015163, "learning_rate": 3.4691841547569883e-06, "loss": 1.3492, "step": 2604 }, { "epoch": 1.7448091091761553, "grad_norm": 5.500794716107556, "learning_rate": 3.4678368038428633e-06, "loss": 1.3543, "step": 2605 }, { "epoch": 1.7454789015405225, "grad_norm": 2.2342536655804937, "learning_rate": 3.466489122173646e-06, "loss": 1.4045, "step": 2606 }, { "epoch": 1.7461486939048894, "grad_norm": 3.088452700435491, "learning_rate": 3.4651411102098985e-06, "loss": 1.211, "step": 2607 }, { "epoch": 1.7468184862692566, "grad_norm": 6.219736828071354, "learning_rate": 3.463792768412303e-06, "loss": 1.3944, "step": 2608 }, { "epoch": 1.7474882786336234, "grad_norm": 2.3515842301203493, "learning_rate": 3.4624440972416478e-06, "loss": 1.3262, "step": 2609 }, { "epoch": 1.7481580709979907, "grad_norm": 2.3360036590936915, "learning_rate": 3.4610950971588395e-06, "loss": 1.1724, "step": 2610 }, { "epoch": 1.7488278633623575, "grad_norm": 2.546339418938239, "learning_rate": 3.4597457686248927e-06, "loss": 1.1989, "step": 2611 }, { "epoch": 1.7494976557267248, "grad_norm": 2.3474016055261355, "learning_rate": 3.458396112100938e-06, "loss": 1.3728, "step": 2612 }, { "epoch": 1.7501674480910918, "grad_norm": 2.1073523498501445, "learning_rate": 3.4570461280482133e-06, "loss": 1.1949, "step": 2613 }, { "epoch": 1.7508372404554589, "grad_norm": 2.1773931258298, "learning_rate": 3.455695816928074e-06, "loss": 1.3807, "step": 2614 }, { "epoch": 1.751507032819826, "grad_norm": 2.4034881751712573, "learning_rate": 3.454345179201983e-06, "loss": 1.2191, "step": 2615 }, { "epoch": 1.752176825184193, "grad_norm": 2.2606653386052247, "learning_rate": 3.4529942153315178e-06, "loss": 1.1829, "step": 2616 }, { "epoch": 1.75284661754856, "grad_norm": 3.22913991713672, "learning_rate": 3.4516429257783634e-06, "loss": 1.1321, "step": 2617 }, { "epoch": 1.753516409912927, "grad_norm": 2.223350606570397, "learning_rate": 3.4502913110043213e-06, "loss": 1.1739, "step": 2618 }, { "epoch": 1.754186202277294, "grad_norm": 3.169595068642921, "learning_rate": 3.4489393714713e-06, "loss": 1.2416, "step": 2619 }, { "epoch": 1.754855994641661, "grad_norm": 2.840999118895771, "learning_rate": 3.4475871076413214e-06, "loss": 1.0134, "step": 2620 }, { "epoch": 1.7555257870060281, "grad_norm": 2.148554371456117, "learning_rate": 3.4462345199765156e-06, "loss": 1.3904, "step": 2621 }, { "epoch": 1.7561955793703952, "grad_norm": 2.339420853434442, "learning_rate": 3.4448816089391266e-06, "loss": 1.2836, "step": 2622 }, { "epoch": 1.7568653717347622, "grad_norm": 2.278319921693922, "learning_rate": 3.443528374991507e-06, "loss": 1.4004, "step": 2623 }, { "epoch": 1.7575351640991292, "grad_norm": 2.7068885415224346, "learning_rate": 3.442174818596121e-06, "loss": 1.053, "step": 2624 }, { "epoch": 1.7582049564634963, "grad_norm": 2.4363845955250625, "learning_rate": 3.44082094021554e-06, "loss": 1.308, "step": 2625 }, { "epoch": 1.7588747488278633, "grad_norm": 3.2913464064872593, "learning_rate": 3.439466740312449e-06, "loss": 1.1522, "step": 2626 }, { "epoch": 1.7595445411922304, "grad_norm": 3.297041961357651, "learning_rate": 3.4381122193496418e-06, "loss": 1.1931, "step": 2627 }, { "epoch": 1.7602143335565974, "grad_norm": 2.2750016137203204, "learning_rate": 3.4367573777900206e-06, "loss": 1.3389, "step": 2628 }, { "epoch": 1.7608841259209647, "grad_norm": 2.4723971168864085, "learning_rate": 3.435402216096598e-06, "loss": 1.1324, "step": 2629 }, { "epoch": 1.7615539182853315, "grad_norm": 2.4621461029074676, "learning_rate": 3.4340467347324967e-06, "loss": 1.2176, "step": 2630 }, { "epoch": 1.7622237106496987, "grad_norm": 2.213297579197219, "learning_rate": 3.4326909341609482e-06, "loss": 1.2437, "step": 2631 }, { "epoch": 1.7628935030140656, "grad_norm": 2.286548915494249, "learning_rate": 3.4313348148452922e-06, "loss": 1.2804, "step": 2632 }, { "epoch": 1.7635632953784328, "grad_norm": 2.262689128435164, "learning_rate": 3.4299783772489794e-06, "loss": 1.3623, "step": 2633 }, { "epoch": 1.7642330877427996, "grad_norm": 2.2539925160803356, "learning_rate": 3.4286216218355663e-06, "loss": 1.3305, "step": 2634 }, { "epoch": 1.764902880107167, "grad_norm": 2.1879786708437203, "learning_rate": 3.4272645490687205e-06, "loss": 1.1826, "step": 2635 }, { "epoch": 1.7655726724715337, "grad_norm": 2.336848885430085, "learning_rate": 3.425907159412218e-06, "loss": 1.4608, "step": 2636 }, { "epoch": 1.766242464835901, "grad_norm": 3.4251722615542364, "learning_rate": 3.424549453329941e-06, "loss": 1.3831, "step": 2637 }, { "epoch": 1.7669122572002678, "grad_norm": 2.3252149336819, "learning_rate": 3.4231914312858817e-06, "loss": 1.2556, "step": 2638 }, { "epoch": 1.767582049564635, "grad_norm": 2.163202808393781, "learning_rate": 3.4218330937441397e-06, "loss": 1.2955, "step": 2639 }, { "epoch": 1.7682518419290019, "grad_norm": 2.344341789187305, "learning_rate": 3.420474441168923e-06, "loss": 1.2465, "step": 2640 }, { "epoch": 1.7689216342933691, "grad_norm": 2.443047616884077, "learning_rate": 3.4191154740245458e-06, "loss": 1.2886, "step": 2641 }, { "epoch": 1.769591426657736, "grad_norm": 2.966833169471203, "learning_rate": 3.41775619277543e-06, "loss": 1.2092, "step": 2642 }, { "epoch": 1.7702612190221032, "grad_norm": 2.9570181366228483, "learning_rate": 3.4163965978861085e-06, "loss": 1.3341, "step": 2643 }, { "epoch": 1.7709310113864702, "grad_norm": 2.9740318191768655, "learning_rate": 3.4150366898212157e-06, "loss": 1.2229, "step": 2644 }, { "epoch": 1.7716008037508373, "grad_norm": 2.210486716450933, "learning_rate": 3.4136764690454972e-06, "loss": 1.4144, "step": 2645 }, { "epoch": 1.7722705961152043, "grad_norm": 2.595334622090435, "learning_rate": 3.412315936023803e-06, "loss": 1.3695, "step": 2646 }, { "epoch": 1.7729403884795714, "grad_norm": 2.3316830922135368, "learning_rate": 3.410955091221092e-06, "loss": 1.3312, "step": 2647 }, { "epoch": 1.7736101808439384, "grad_norm": 3.6223303174429704, "learning_rate": 3.4095939351024275e-06, "loss": 1.083, "step": 2648 }, { "epoch": 1.7742799732083054, "grad_norm": 6.167130708836196, "learning_rate": 3.4082324681329815e-06, "loss": 1.1542, "step": 2649 }, { "epoch": 1.7749497655726725, "grad_norm": 2.179179946882595, "learning_rate": 3.4068706907780287e-06, "loss": 1.2947, "step": 2650 }, { "epoch": 1.7756195579370395, "grad_norm": 2.7593122040758993, "learning_rate": 3.4055086035029544e-06, "loss": 1.3599, "step": 2651 }, { "epoch": 1.7762893503014066, "grad_norm": 2.612004421131549, "learning_rate": 3.404146206773246e-06, "loss": 1.3783, "step": 2652 }, { "epoch": 1.7769591426657736, "grad_norm": 4.394671932864723, "learning_rate": 3.4027835010545e-06, "loss": 1.3068, "step": 2653 }, { "epoch": 1.7776289350301406, "grad_norm": 2.0831789000186323, "learning_rate": 3.4014204868124136e-06, "loss": 1.1868, "step": 2654 }, { "epoch": 1.7782987273945077, "grad_norm": 2.646401232892572, "learning_rate": 3.4000571645127958e-06, "loss": 1.1631, "step": 2655 }, { "epoch": 1.7789685197588747, "grad_norm": 4.244471372463427, "learning_rate": 3.398693534621555e-06, "loss": 1.1772, "step": 2656 }, { "epoch": 1.7796383121232418, "grad_norm": 3.774480581532351, "learning_rate": 3.3973295976047093e-06, "loss": 1.201, "step": 2657 }, { "epoch": 1.7803081044876088, "grad_norm": 2.576956389617284, "learning_rate": 3.3959653539283775e-06, "loss": 1.3469, "step": 2658 }, { "epoch": 1.7809778968519758, "grad_norm": 3.0199545825673373, "learning_rate": 3.3946008040587883e-06, "loss": 1.4066, "step": 2659 }, { "epoch": 1.781647689216343, "grad_norm": 2.2261030822414876, "learning_rate": 3.3932359484622703e-06, "loss": 1.3189, "step": 2660 }, { "epoch": 1.78231748158071, "grad_norm": 2.447821404781074, "learning_rate": 3.3918707876052594e-06, "loss": 1.2182, "step": 2661 }, { "epoch": 1.7829872739450772, "grad_norm": 2.540127085618783, "learning_rate": 3.390505321954293e-06, "loss": 1.139, "step": 2662 }, { "epoch": 1.783657066309444, "grad_norm": 2.615292771671534, "learning_rate": 3.3891395519760173e-06, "loss": 1.3403, "step": 2663 }, { "epoch": 1.7843268586738112, "grad_norm": 2.9737474377860273, "learning_rate": 3.3877734781371775e-06, "loss": 1.1801, "step": 2664 }, { "epoch": 1.784996651038178, "grad_norm": 2.2938644167053543, "learning_rate": 3.3864071009046263e-06, "loss": 1.4269, "step": 2665 }, { "epoch": 1.7856664434025453, "grad_norm": 2.7275047952594393, "learning_rate": 3.385040420745318e-06, "loss": 1.1135, "step": 2666 }, { "epoch": 1.7863362357669121, "grad_norm": 2.2598755143199525, "learning_rate": 3.3836734381263116e-06, "loss": 1.1119, "step": 2667 }, { "epoch": 1.7870060281312794, "grad_norm": 2.741783510547853, "learning_rate": 3.382306153514768e-06, "loss": 1.3721, "step": 2668 }, { "epoch": 1.7876758204956462, "grad_norm": 4.350932020270433, "learning_rate": 3.3809385673779545e-06, "loss": 1.0851, "step": 2669 }, { "epoch": 1.7883456128600135, "grad_norm": 5.143553312882983, "learning_rate": 3.3795706801832367e-06, "loss": 1.2116, "step": 2670 }, { "epoch": 1.7890154052243803, "grad_norm": 2.2579595137712394, "learning_rate": 3.378202492398087e-06, "loss": 1.3191, "step": 2671 }, { "epoch": 1.7896851975887476, "grad_norm": 3.8751633212134404, "learning_rate": 3.376834004490079e-06, "loss": 1.2405, "step": 2672 }, { "epoch": 1.7903549899531144, "grad_norm": 3.674076798131482, "learning_rate": 3.375465216926889e-06, "loss": 1.1766, "step": 2673 }, { "epoch": 1.7910247823174816, "grad_norm": 2.810870988625837, "learning_rate": 3.374096130176295e-06, "loss": 1.351, "step": 2674 }, { "epoch": 1.7916945746818487, "grad_norm": 2.229385233263155, "learning_rate": 3.3727267447061785e-06, "loss": 1.3297, "step": 2675 }, { "epoch": 1.7923643670462157, "grad_norm": 2.3777761062760954, "learning_rate": 3.371357060984523e-06, "loss": 1.2761, "step": 2676 }, { "epoch": 1.7930341594105828, "grad_norm": 2.2444155514857282, "learning_rate": 3.369987079479413e-06, "loss": 1.2744, "step": 2677 }, { "epoch": 1.7937039517749498, "grad_norm": 3.4790921709738085, "learning_rate": 3.3686168006590346e-06, "loss": 1.0194, "step": 2678 }, { "epoch": 1.7943737441393168, "grad_norm": 1.9668868494642864, "learning_rate": 3.3672462249916783e-06, "loss": 1.16, "step": 2679 }, { "epoch": 1.7950435365036839, "grad_norm": 2.5311156271002027, "learning_rate": 3.365875352945731e-06, "loss": 1.2403, "step": 2680 }, { "epoch": 1.795713328868051, "grad_norm": 2.7051229396906074, "learning_rate": 3.3645041849896852e-06, "loss": 1.2891, "step": 2681 }, { "epoch": 1.796383121232418, "grad_norm": 2.439349209820296, "learning_rate": 3.363132721592135e-06, "loss": 1.3715, "step": 2682 }, { "epoch": 1.797052913596785, "grad_norm": 4.990054607230462, "learning_rate": 3.3617609632217694e-06, "loss": 1.2434, "step": 2683 }, { "epoch": 1.797722705961152, "grad_norm": 2.9538458583267304, "learning_rate": 3.360388910347387e-06, "loss": 1.254, "step": 2684 }, { "epoch": 1.798392498325519, "grad_norm": 2.1200233578419643, "learning_rate": 3.359016563437878e-06, "loss": 1.4015, "step": 2685 }, { "epoch": 1.799062290689886, "grad_norm": 4.253076918976518, "learning_rate": 3.3576439229622414e-06, "loss": 1.3309, "step": 2686 }, { "epoch": 1.7997320830542531, "grad_norm": 2.964549831216243, "learning_rate": 3.3562709893895696e-06, "loss": 1.1501, "step": 2687 }, { "epoch": 1.8004018754186202, "grad_norm": 2.5820953980373202, "learning_rate": 3.35489776318906e-06, "loss": 1.3288, "step": 2688 }, { "epoch": 1.8010716677829874, "grad_norm": 2.1222225493923186, "learning_rate": 3.3535242448300076e-06, "loss": 1.2059, "step": 2689 }, { "epoch": 1.8017414601473543, "grad_norm": 2.564809146468308, "learning_rate": 3.3521504347818083e-06, "loss": 1.2359, "step": 2690 }, { "epoch": 1.8024112525117215, "grad_norm": 2.068300566608026, "learning_rate": 3.350776333513956e-06, "loss": 1.3512, "step": 2691 }, { "epoch": 1.8030810448760883, "grad_norm": 3.6882858279349375, "learning_rate": 3.3494019414960477e-06, "loss": 1.2111, "step": 2692 }, { "epoch": 1.8037508372404556, "grad_norm": 2.358509427016276, "learning_rate": 3.348027259197774e-06, "loss": 1.1307, "step": 2693 }, { "epoch": 1.8044206296048224, "grad_norm": 2.141208891639953, "learning_rate": 3.3466522870889316e-06, "loss": 1.3946, "step": 2694 }, { "epoch": 1.8050904219691897, "grad_norm": 3.9103729240171115, "learning_rate": 3.3452770256394107e-06, "loss": 1.1511, "step": 2695 }, { "epoch": 1.8057602143335565, "grad_norm": 3.0290253247975953, "learning_rate": 3.3439014753192033e-06, "loss": 1.3022, "step": 2696 }, { "epoch": 1.8064300066979238, "grad_norm": 3.205636714501436, "learning_rate": 3.3425256365983983e-06, "loss": 1.1923, "step": 2697 }, { "epoch": 1.8070997990622906, "grad_norm": 2.1700640230036297, "learning_rate": 3.3411495099471857e-06, "loss": 1.3415, "step": 2698 }, { "epoch": 1.8077695914266578, "grad_norm": 2.3348794832098836, "learning_rate": 3.339773095835851e-06, "loss": 1.4349, "step": 2699 }, { "epoch": 1.8084393837910246, "grad_norm": 2.6205104867005447, "learning_rate": 3.3383963947347803e-06, "loss": 1.2852, "step": 2700 }, { "epoch": 1.809109176155392, "grad_norm": 2.578481773899073, "learning_rate": 3.3370194071144547e-06, "loss": 1.3648, "step": 2701 }, { "epoch": 1.8097789685197587, "grad_norm": 2.3388185624468485, "learning_rate": 3.335642133445458e-06, "loss": 1.3129, "step": 2702 }, { "epoch": 1.810448760884126, "grad_norm": 3.118298671985833, "learning_rate": 3.334264574198467e-06, "loss": 1.2792, "step": 2703 }, { "epoch": 1.8111185532484928, "grad_norm": 2.1879399265615946, "learning_rate": 3.3328867298442597e-06, "loss": 1.3493, "step": 2704 }, { "epoch": 1.81178834561286, "grad_norm": 2.558994608484323, "learning_rate": 3.331508600853708e-06, "loss": 1.1626, "step": 2705 }, { "epoch": 1.812458137977227, "grad_norm": 2.4611726744950855, "learning_rate": 3.330130187697785e-06, "loss": 1.4817, "step": 2706 }, { "epoch": 1.8131279303415941, "grad_norm": 2.4522177964587315, "learning_rate": 3.328751490847557e-06, "loss": 1.3749, "step": 2707 }, { "epoch": 1.8137977227059612, "grad_norm": 2.3474075570827204, "learning_rate": 3.327372510774191e-06, "loss": 1.3103, "step": 2708 }, { "epoch": 1.8144675150703282, "grad_norm": 2.2692201508608623, "learning_rate": 3.3259932479489477e-06, "loss": 1.2724, "step": 2709 }, { "epoch": 1.8151373074346953, "grad_norm": 2.0921539941900575, "learning_rate": 3.3246137028431864e-06, "loss": 1.3927, "step": 2710 }, { "epoch": 1.8158070997990623, "grad_norm": 2.1292982906287, "learning_rate": 3.323233875928362e-06, "loss": 1.1204, "step": 2711 }, { "epoch": 1.8164768921634293, "grad_norm": 3.552239986201101, "learning_rate": 3.3218537676760255e-06, "loss": 1.1418, "step": 2712 }, { "epoch": 1.8171466845277964, "grad_norm": 2.711186103798218, "learning_rate": 3.3204733785578254e-06, "loss": 1.2585, "step": 2713 }, { "epoch": 1.8178164768921634, "grad_norm": 2.20834852565053, "learning_rate": 3.3190927090455046e-06, "loss": 1.3048, "step": 2714 }, { "epoch": 1.8184862692565305, "grad_norm": 2.3611317211051768, "learning_rate": 3.317711759610902e-06, "loss": 1.3339, "step": 2715 }, { "epoch": 1.8191560616208975, "grad_norm": 2.683981010637687, "learning_rate": 3.316330530725953e-06, "loss": 1.3025, "step": 2716 }, { "epoch": 1.8198258539852645, "grad_norm": 2.082075091160055, "learning_rate": 3.3149490228626886e-06, "loss": 1.3572, "step": 2717 }, { "epoch": 1.8204956463496316, "grad_norm": 2.91645009261138, "learning_rate": 3.3135672364932337e-06, "loss": 1.2206, "step": 2718 }, { "epoch": 1.8211654387139986, "grad_norm": 2.2490584777646787, "learning_rate": 3.3121851720898094e-06, "loss": 1.424, "step": 2719 }, { "epoch": 1.8218352310783659, "grad_norm": 2.3566416789139266, "learning_rate": 3.310802830124733e-06, "loss": 1.1989, "step": 2720 }, { "epoch": 1.8225050234427327, "grad_norm": 2.6674559207382185, "learning_rate": 3.3094202110704145e-06, "loss": 1.3058, "step": 2721 }, { "epoch": 1.8231748158071, "grad_norm": 2.446932345525975, "learning_rate": 3.308037315399359e-06, "loss": 1.2794, "step": 2722 }, { "epoch": 1.8238446081714668, "grad_norm": 2.10088476803347, "learning_rate": 3.306654143584167e-06, "loss": 1.0642, "step": 2723 }, { "epoch": 1.824514400535834, "grad_norm": 4.381495994263629, "learning_rate": 3.3052706960975332e-06, "loss": 1.3301, "step": 2724 }, { "epoch": 1.8251841929002008, "grad_norm": 3.980407361578165, "learning_rate": 3.3038869734122466e-06, "loss": 1.3575, "step": 2725 }, { "epoch": 1.825853985264568, "grad_norm": 2.269658973187834, "learning_rate": 3.3025029760011883e-06, "loss": 1.2901, "step": 2726 }, { "epoch": 1.826523777628935, "grad_norm": 4.07716770003165, "learning_rate": 3.3011187043373378e-06, "loss": 1.3917, "step": 2727 }, { "epoch": 1.8271935699933022, "grad_norm": 2.886120061091163, "learning_rate": 3.2997341588937624e-06, "loss": 1.2494, "step": 2728 }, { "epoch": 1.827863362357669, "grad_norm": 2.2171291261868817, "learning_rate": 3.298349340143629e-06, "loss": 1.3287, "step": 2729 }, { "epoch": 1.8285331547220363, "grad_norm": 2.1623234591186677, "learning_rate": 3.2969642485601923e-06, "loss": 1.4616, "step": 2730 }, { "epoch": 1.829202947086403, "grad_norm": 2.3859594955583323, "learning_rate": 3.2955788846168046e-06, "loss": 1.1784, "step": 2731 }, { "epoch": 1.8298727394507703, "grad_norm": 2.1879476284003996, "learning_rate": 3.294193248786908e-06, "loss": 1.2302, "step": 2732 }, { "epoch": 1.8305425318151372, "grad_norm": 2.4925780906016315, "learning_rate": 3.292807341544041e-06, "loss": 1.1027, "step": 2733 }, { "epoch": 1.8312123241795044, "grad_norm": 2.4532784412786506, "learning_rate": 3.291421163361831e-06, "loss": 1.2983, "step": 2734 }, { "epoch": 1.8318821165438715, "grad_norm": 2.775055638131861, "learning_rate": 3.2900347147140015e-06, "loss": 1.3295, "step": 2735 }, { "epoch": 1.8325519089082385, "grad_norm": 2.183201419507375, "learning_rate": 3.2886479960743655e-06, "loss": 1.0292, "step": 2736 }, { "epoch": 1.8332217012726055, "grad_norm": 3.583150356914991, "learning_rate": 3.287261007916831e-06, "loss": 1.176, "step": 2737 }, { "epoch": 1.8338914936369726, "grad_norm": 3.128980750498469, "learning_rate": 3.2858737507153947e-06, "loss": 1.4713, "step": 2738 }, { "epoch": 1.8345612860013396, "grad_norm": 2.307541190883782, "learning_rate": 3.2844862249441496e-06, "loss": 1.2939, "step": 2739 }, { "epoch": 1.8352310783657066, "grad_norm": 2.21312318903937, "learning_rate": 3.2830984310772764e-06, "loss": 1.2172, "step": 2740 }, { "epoch": 1.8359008707300737, "grad_norm": 2.495090811940208, "learning_rate": 3.28171036958905e-06, "loss": 1.1751, "step": 2741 }, { "epoch": 1.8365706630944407, "grad_norm": 3.9682877002254098, "learning_rate": 3.280322040953835e-06, "loss": 1.4487, "step": 2742 }, { "epoch": 1.8372404554588078, "grad_norm": 2.879496029362145, "learning_rate": 3.27893344564609e-06, "loss": 1.1475, "step": 2743 }, { "epoch": 1.8379102478231748, "grad_norm": 2.0769504850448106, "learning_rate": 3.277544584140361e-06, "loss": 1.2917, "step": 2744 }, { "epoch": 1.8385800401875418, "grad_norm": 4.027110796997352, "learning_rate": 3.2761554569112887e-06, "loss": 1.218, "step": 2745 }, { "epoch": 1.8392498325519089, "grad_norm": 2.2614041947339274, "learning_rate": 3.274766064433601e-06, "loss": 1.378, "step": 2746 }, { "epoch": 1.839919624916276, "grad_norm": 2.4369897494358894, "learning_rate": 3.27337640718212e-06, "loss": 1.1597, "step": 2747 }, { "epoch": 1.840589417280643, "grad_norm": 2.1036537172684033, "learning_rate": 3.271986485631755e-06, "loss": 1.1627, "step": 2748 }, { "epoch": 1.84125920964501, "grad_norm": 2.1413603302512327, "learning_rate": 3.270596300257509e-06, "loss": 1.2618, "step": 2749 }, { "epoch": 1.841929002009377, "grad_norm": 2.6188988325722797, "learning_rate": 3.269205851534471e-06, "loss": 1.3257, "step": 2750 }, { "epoch": 1.8425987943737443, "grad_norm": 2.1115187608554162, "learning_rate": 3.2678151399378245e-06, "loss": 1.2042, "step": 2751 }, { "epoch": 1.8432685867381111, "grad_norm": 3.5037517674183962, "learning_rate": 3.2664241659428396e-06, "loss": 1.366, "step": 2752 }, { "epoch": 1.8439383791024784, "grad_norm": 3.746499991615757, "learning_rate": 3.2650329300248772e-06, "loss": 1.2705, "step": 2753 }, { "epoch": 1.8446081714668452, "grad_norm": 2.078927289238101, "learning_rate": 3.2636414326593884e-06, "loss": 1.2743, "step": 2754 }, { "epoch": 1.8452779638312125, "grad_norm": 2.3625522560811367, "learning_rate": 3.262249674321912e-06, "loss": 1.3906, "step": 2755 }, { "epoch": 1.8459477561955793, "grad_norm": 4.289541180322981, "learning_rate": 3.2608576554880775e-06, "loss": 1.3887, "step": 2756 }, { "epoch": 1.8466175485599465, "grad_norm": 2.768924663523329, "learning_rate": 3.2594653766336027e-06, "loss": 1.2731, "step": 2757 }, { "epoch": 1.8472873409243133, "grad_norm": 2.2115950196549328, "learning_rate": 3.258072838234294e-06, "loss": 1.3209, "step": 2758 }, { "epoch": 1.8479571332886806, "grad_norm": 2.067556951899147, "learning_rate": 3.256680040766048e-06, "loss": 1.1112, "step": 2759 }, { "epoch": 1.8486269256530474, "grad_norm": 2.2663385176393565, "learning_rate": 3.255286984704847e-06, "loss": 1.2114, "step": 2760 }, { "epoch": 1.8492967180174147, "grad_norm": 2.449473728302127, "learning_rate": 3.2538936705267643e-06, "loss": 1.2546, "step": 2761 }, { "epoch": 1.8499665103817815, "grad_norm": 2.3836694565795553, "learning_rate": 3.2525000987079615e-06, "loss": 1.2356, "step": 2762 }, { "epoch": 1.8506363027461488, "grad_norm": 4.376262504310592, "learning_rate": 3.2511062697246853e-06, "loss": 1.3025, "step": 2763 }, { "epoch": 1.8513060951105156, "grad_norm": 2.2076212836772116, "learning_rate": 3.2497121840532734e-06, "loss": 1.2596, "step": 2764 }, { "epoch": 1.8519758874748828, "grad_norm": 2.2525590670250852, "learning_rate": 3.2483178421701507e-06, "loss": 1.3617, "step": 2765 }, { "epoch": 1.8526456798392499, "grad_norm": 2.1228568853771552, "learning_rate": 3.2469232445518274e-06, "loss": 1.2024, "step": 2766 }, { "epoch": 1.853315472203617, "grad_norm": 2.2619360820995364, "learning_rate": 3.2455283916749034e-06, "loss": 1.3875, "step": 2767 }, { "epoch": 1.853985264567984, "grad_norm": 2.3483971913911637, "learning_rate": 3.2441332840160665e-06, "loss": 1.4013, "step": 2768 }, { "epoch": 1.854655056932351, "grad_norm": 2.473627957402622, "learning_rate": 3.2427379220520883e-06, "loss": 1.2027, "step": 2769 }, { "epoch": 1.855324849296718, "grad_norm": 2.157584051707231, "learning_rate": 3.2413423062598296e-06, "loss": 1.2891, "step": 2770 }, { "epoch": 1.855994641661085, "grad_norm": 2.1100017590095295, "learning_rate": 3.2399464371162382e-06, "loss": 1.2528, "step": 2771 }, { "epoch": 1.8566644340254521, "grad_norm": 2.1158890320628805, "learning_rate": 3.238550315098348e-06, "loss": 1.3645, "step": 2772 }, { "epoch": 1.8573342263898192, "grad_norm": 3.9895305737183455, "learning_rate": 3.237153940683278e-06, "loss": 1.2697, "step": 2773 }, { "epoch": 1.8580040187541862, "grad_norm": 2.597680641826693, "learning_rate": 3.2357573143482364e-06, "loss": 1.1877, "step": 2774 }, { "epoch": 1.8586738111185532, "grad_norm": 2.1039575171179306, "learning_rate": 3.234360436570514e-06, "loss": 1.2404, "step": 2775 }, { "epoch": 1.8593436034829203, "grad_norm": 3.076064321460503, "learning_rate": 3.2329633078274904e-06, "loss": 1.2177, "step": 2776 }, { "epoch": 1.8600133958472873, "grad_norm": 2.50060787313977, "learning_rate": 3.231565928596629e-06, "loss": 1.1723, "step": 2777 }, { "epoch": 1.8606831882116543, "grad_norm": 4.572679874985365, "learning_rate": 3.2301682993554806e-06, "loss": 1.305, "step": 2778 }, { "epoch": 1.8613529805760214, "grad_norm": 3.4435634700861946, "learning_rate": 3.2287704205816793e-06, "loss": 1.1773, "step": 2779 }, { "epoch": 1.8620227729403884, "grad_norm": 4.205836546194661, "learning_rate": 3.227372292752947e-06, "loss": 1.2875, "step": 2780 }, { "epoch": 1.8626925653047555, "grad_norm": 2.4113140318971, "learning_rate": 3.2259739163470878e-06, "loss": 1.2393, "step": 2781 }, { "epoch": 1.8633623576691227, "grad_norm": 2.1310711208384725, "learning_rate": 3.2245752918419942e-06, "loss": 1.0612, "step": 2782 }, { "epoch": 1.8640321500334895, "grad_norm": 2.3671612219286557, "learning_rate": 3.223176419715639e-06, "loss": 1.4673, "step": 2783 }, { "epoch": 1.8647019423978568, "grad_norm": 3.270964483646576, "learning_rate": 3.221777300446085e-06, "loss": 1.2705, "step": 2784 }, { "epoch": 1.8653717347622236, "grad_norm": 3.0816390136142875, "learning_rate": 3.220377934511475e-06, "loss": 1.111, "step": 2785 }, { "epoch": 1.8660415271265909, "grad_norm": 2.2859498417925006, "learning_rate": 3.2189783223900383e-06, "loss": 1.3442, "step": 2786 }, { "epoch": 1.8667113194909577, "grad_norm": 2.2534999466018513, "learning_rate": 3.2175784645600884e-06, "loss": 1.247, "step": 2787 }, { "epoch": 1.867381111855325, "grad_norm": 4.43337462083418, "learning_rate": 3.2161783615000205e-06, "loss": 1.2069, "step": 2788 }, { "epoch": 1.8680509042196918, "grad_norm": 3.4559831566974712, "learning_rate": 3.2147780136883166e-06, "loss": 1.3444, "step": 2789 }, { "epoch": 1.868720696584059, "grad_norm": 4.087645652311553, "learning_rate": 3.2133774216035408e-06, "loss": 1.1454, "step": 2790 }, { "epoch": 1.8693904889484259, "grad_norm": 3.0965370403948222, "learning_rate": 3.2119765857243406e-06, "loss": 1.1145, "step": 2791 }, { "epoch": 1.8700602813127931, "grad_norm": 2.452210775884313, "learning_rate": 3.210575506529448e-06, "loss": 1.4465, "step": 2792 }, { "epoch": 1.87073007367716, "grad_norm": 2.2831988937352006, "learning_rate": 3.209174184497676e-06, "loss": 0.9735, "step": 2793 }, { "epoch": 1.8713998660415272, "grad_norm": 4.063765204092155, "learning_rate": 3.207772620107923e-06, "loss": 1.3654, "step": 2794 }, { "epoch": 1.872069658405894, "grad_norm": 2.3178999747238307, "learning_rate": 3.206370813839168e-06, "loss": 1.3166, "step": 2795 }, { "epoch": 1.8727394507702613, "grad_norm": 2.8996209560630297, "learning_rate": 3.204968766170475e-06, "loss": 1.0862, "step": 2796 }, { "epoch": 1.8734092431346283, "grad_norm": 2.2186685999390607, "learning_rate": 3.2035664775809883e-06, "loss": 1.2533, "step": 2797 }, { "epoch": 1.8740790354989953, "grad_norm": 3.273312248076836, "learning_rate": 3.202163948549936e-06, "loss": 1.1909, "step": 2798 }, { "epoch": 1.8747488278633624, "grad_norm": 2.401304047227551, "learning_rate": 3.2007611795566275e-06, "loss": 1.3912, "step": 2799 }, { "epoch": 1.8754186202277294, "grad_norm": 2.533699821668855, "learning_rate": 3.1993581710804544e-06, "loss": 1.2286, "step": 2800 }, { "epoch": 1.8760884125920965, "grad_norm": 3.751857866477904, "learning_rate": 3.1979549236008905e-06, "loss": 1.3707, "step": 2801 }, { "epoch": 1.8767582049564635, "grad_norm": 3.147764359884369, "learning_rate": 3.1965514375974915e-06, "loss": 1.1856, "step": 2802 }, { "epoch": 1.8774279973208305, "grad_norm": 2.91705528210309, "learning_rate": 3.195147713549894e-06, "loss": 1.2012, "step": 2803 }, { "epoch": 1.8780977896851976, "grad_norm": 2.680599429275311, "learning_rate": 3.1937437519378163e-06, "loss": 1.0668, "step": 2804 }, { "epoch": 1.8787675820495646, "grad_norm": 2.203284223521185, "learning_rate": 3.1923395532410572e-06, "loss": 1.2467, "step": 2805 }, { "epoch": 1.8794373744139317, "grad_norm": 2.736246950973001, "learning_rate": 3.1909351179394972e-06, "loss": 1.2333, "step": 2806 }, { "epoch": 1.8801071667782987, "grad_norm": 3.2179059388591797, "learning_rate": 3.1895304465130976e-06, "loss": 1.325, "step": 2807 }, { "epoch": 1.8807769591426657, "grad_norm": 3.2932115093375938, "learning_rate": 3.188125539441901e-06, "loss": 1.0636, "step": 2808 }, { "epoch": 1.8814467515070328, "grad_norm": 2.4325376038767024, "learning_rate": 3.1867203972060286e-06, "loss": 1.2253, "step": 2809 }, { "epoch": 1.8821165438713998, "grad_norm": 3.2545018615412205, "learning_rate": 3.1853150202856835e-06, "loss": 1.0687, "step": 2810 }, { "epoch": 1.882786336235767, "grad_norm": 2.507286064327203, "learning_rate": 3.1839094091611496e-06, "loss": 1.3504, "step": 2811 }, { "epoch": 1.883456128600134, "grad_norm": 2.309412934008655, "learning_rate": 3.1825035643127888e-06, "loss": 1.3141, "step": 2812 }, { "epoch": 1.8841259209645012, "grad_norm": 2.9221125936404433, "learning_rate": 3.1810974862210454e-06, "loss": 1.2434, "step": 2813 }, { "epoch": 1.884795713328868, "grad_norm": 2.428043435159344, "learning_rate": 3.1796911753664405e-06, "loss": 1.2866, "step": 2814 }, { "epoch": 1.8854655056932352, "grad_norm": 4.851193883712991, "learning_rate": 3.178284632229578e-06, "loss": 1.2244, "step": 2815 }, { "epoch": 1.886135298057602, "grad_norm": 3.7735680365114868, "learning_rate": 3.1768778572911376e-06, "loss": 1.218, "step": 2816 }, { "epoch": 1.8868050904219693, "grad_norm": 3.389749205284753, "learning_rate": 3.1754708510318812e-06, "loss": 1.2717, "step": 2817 }, { "epoch": 1.8874748827863361, "grad_norm": 2.24453804999996, "learning_rate": 3.174063613932648e-06, "loss": 1.2556, "step": 2818 }, { "epoch": 1.8881446751507034, "grad_norm": 3.216840377485859, "learning_rate": 3.1726561464743584e-06, "loss": 1.1674, "step": 2819 }, { "epoch": 1.8888144675150702, "grad_norm": 2.3024688800097364, "learning_rate": 3.1712484491380068e-06, "loss": 1.3174, "step": 2820 }, { "epoch": 1.8894842598794375, "grad_norm": 3.5420850973197577, "learning_rate": 3.1698405224046725e-06, "loss": 1.2521, "step": 2821 }, { "epoch": 1.8901540522438043, "grad_norm": 2.604162785476245, "learning_rate": 3.168432366755507e-06, "loss": 1.0762, "step": 2822 }, { "epoch": 1.8908238446081715, "grad_norm": 2.0945647964605394, "learning_rate": 3.1670239826717453e-06, "loss": 1.3079, "step": 2823 }, { "epoch": 1.8914936369725384, "grad_norm": 4.105268246983148, "learning_rate": 3.165615370634697e-06, "loss": 1.4, "step": 2824 }, { "epoch": 1.8921634293369056, "grad_norm": 2.5271089492019163, "learning_rate": 3.1642065311257507e-06, "loss": 1.0675, "step": 2825 }, { "epoch": 1.8928332217012724, "grad_norm": 2.1460266499340004, "learning_rate": 3.162797464626373e-06, "loss": 1.1452, "step": 2826 }, { "epoch": 1.8935030140656397, "grad_norm": 2.53402573929184, "learning_rate": 3.1613881716181078e-06, "loss": 1.2094, "step": 2827 }, { "epoch": 1.8941728064300067, "grad_norm": 2.411046578799199, "learning_rate": 3.159978652582576e-06, "loss": 1.1369, "step": 2828 }, { "epoch": 1.8948425987943738, "grad_norm": 3.307072718512101, "learning_rate": 3.1585689080014764e-06, "loss": 1.2416, "step": 2829 }, { "epoch": 1.8955123911587408, "grad_norm": 2.150329350890106, "learning_rate": 3.1571589383565847e-06, "loss": 1.2491, "step": 2830 }, { "epoch": 1.8961821835231079, "grad_norm": 2.554159528656219, "learning_rate": 3.155748744129754e-06, "loss": 1.0927, "step": 2831 }, { "epoch": 1.896851975887475, "grad_norm": 2.393607603773714, "learning_rate": 3.1543383258029127e-06, "loss": 1.2623, "step": 2832 }, { "epoch": 1.897521768251842, "grad_norm": 2.6123825519041692, "learning_rate": 3.152927683858067e-06, "loss": 1.2239, "step": 2833 }, { "epoch": 1.898191560616209, "grad_norm": 3.5969338111458695, "learning_rate": 3.1515168187772993e-06, "loss": 1.3286, "step": 2834 }, { "epoch": 1.898861352980576, "grad_norm": 2.229886663865643, "learning_rate": 3.150105731042768e-06, "loss": 1.2528, "step": 2835 }, { "epoch": 1.899531145344943, "grad_norm": 3.0410565416004935, "learning_rate": 3.148694421136708e-06, "loss": 1.3158, "step": 2836 }, { "epoch": 1.90020093770931, "grad_norm": 2.2243602900539305, "learning_rate": 3.1472828895414302e-06, "loss": 1.3443, "step": 2837 }, { "epoch": 1.9008707300736771, "grad_norm": 2.1530017878788628, "learning_rate": 3.14587113673932e-06, "loss": 1.263, "step": 2838 }, { "epoch": 1.9015405224380442, "grad_norm": 3.1733772749412177, "learning_rate": 3.1444591632128407e-06, "loss": 1.409, "step": 2839 }, { "epoch": 1.9022103148024112, "grad_norm": 2.5536606369334445, "learning_rate": 3.1430469694445286e-06, "loss": 1.1596, "step": 2840 }, { "epoch": 1.9028801071667782, "grad_norm": 3.559655612009412, "learning_rate": 3.1416345559169964e-06, "loss": 1.2276, "step": 2841 }, { "epoch": 1.9035498995311455, "grad_norm": 2.180667853542846, "learning_rate": 3.140221923112932e-06, "loss": 1.2695, "step": 2842 }, { "epoch": 1.9042196918955123, "grad_norm": 3.2475859062318726, "learning_rate": 3.138809071515099e-06, "loss": 1.1972, "step": 2843 }, { "epoch": 1.9048894842598796, "grad_norm": 2.3141210023066847, "learning_rate": 3.1373960016063333e-06, "loss": 1.1962, "step": 2844 }, { "epoch": 1.9055592766242464, "grad_norm": 3.3293161623146603, "learning_rate": 3.1359827138695475e-06, "loss": 1.3498, "step": 2845 }, { "epoch": 1.9062290689886137, "grad_norm": 2.232134537734001, "learning_rate": 3.134569208787729e-06, "loss": 1.0605, "step": 2846 }, { "epoch": 1.9068988613529805, "grad_norm": 7.168312609614958, "learning_rate": 3.1331554868439375e-06, "loss": 1.2638, "step": 2847 }, { "epoch": 1.9075686537173477, "grad_norm": 2.2454127317571047, "learning_rate": 3.131741548521309e-06, "loss": 1.3272, "step": 2848 }, { "epoch": 1.9082384460817146, "grad_norm": 2.2309262201620137, "learning_rate": 3.130327394303051e-06, "loss": 1.283, "step": 2849 }, { "epoch": 1.9089082384460818, "grad_norm": 2.0795796452556075, "learning_rate": 3.1289130246724464e-06, "loss": 1.3538, "step": 2850 }, { "epoch": 1.9095780308104486, "grad_norm": 2.252181537988207, "learning_rate": 3.1274984401128527e-06, "loss": 1.3223, "step": 2851 }, { "epoch": 1.910247823174816, "grad_norm": 1.963849923735474, "learning_rate": 3.126083641107698e-06, "loss": 1.1493, "step": 2852 }, { "epoch": 1.9109176155391827, "grad_norm": 3.2296747105600274, "learning_rate": 3.1246686281404864e-06, "loss": 1.4114, "step": 2853 }, { "epoch": 1.91158740790355, "grad_norm": 2.172691035017571, "learning_rate": 3.1232534016947937e-06, "loss": 1.1828, "step": 2854 }, { "epoch": 1.9122572002679168, "grad_norm": 2.580784472200776, "learning_rate": 3.1218379622542685e-06, "loss": 1.2107, "step": 2855 }, { "epoch": 1.912926992632284, "grad_norm": 2.9303465773117554, "learning_rate": 3.1204223103026326e-06, "loss": 1.419, "step": 2856 }, { "epoch": 1.913596784996651, "grad_norm": 2.562101931120855, "learning_rate": 3.1190064463236818e-06, "loss": 1.1166, "step": 2857 }, { "epoch": 1.9142665773610181, "grad_norm": 2.503766529738847, "learning_rate": 3.1175903708012813e-06, "loss": 1.3388, "step": 2858 }, { "epoch": 1.9149363697253852, "grad_norm": 2.1310327139570804, "learning_rate": 3.116174084219372e-06, "loss": 1.4663, "step": 2859 }, { "epoch": 1.9156061620897522, "grad_norm": 3.098800735623731, "learning_rate": 3.1147575870619632e-06, "loss": 1.2499, "step": 2860 }, { "epoch": 1.9162759544541192, "grad_norm": 2.13589436521789, "learning_rate": 3.1133408798131403e-06, "loss": 1.2229, "step": 2861 }, { "epoch": 1.9169457468184863, "grad_norm": 2.284545747746093, "learning_rate": 3.1119239629570576e-06, "loss": 1.2946, "step": 2862 }, { "epoch": 1.9176155391828533, "grad_norm": 4.7616430749319845, "learning_rate": 3.1105068369779414e-06, "loss": 1.272, "step": 2863 }, { "epoch": 1.9182853315472204, "grad_norm": 2.1233585559878176, "learning_rate": 3.1090895023600896e-06, "loss": 1.2661, "step": 2864 }, { "epoch": 1.9189551239115874, "grad_norm": 2.8382729006147076, "learning_rate": 3.1076719595878734e-06, "loss": 1.2518, "step": 2865 }, { "epoch": 1.9196249162759544, "grad_norm": 2.1338799692340884, "learning_rate": 3.106254209145732e-06, "loss": 1.1199, "step": 2866 }, { "epoch": 1.9202947086403215, "grad_norm": 2.6709702441091068, "learning_rate": 3.1048362515181777e-06, "loss": 1.2637, "step": 2867 }, { "epoch": 1.9209645010046885, "grad_norm": 2.3352922288230418, "learning_rate": 3.103418087189793e-06, "loss": 1.2837, "step": 2868 }, { "epoch": 1.9216342933690556, "grad_norm": 3.942844635670942, "learning_rate": 3.10199971664523e-06, "loss": 1.1132, "step": 2869 }, { "epoch": 1.9223040857334226, "grad_norm": 2.363263148880018, "learning_rate": 3.1005811403692134e-06, "loss": 1.2765, "step": 2870 }, { "epoch": 1.9229738780977896, "grad_norm": 3.4789329324246627, "learning_rate": 3.0991623588465365e-06, "loss": 1.2581, "step": 2871 }, { "epoch": 1.9236436704621567, "grad_norm": 2.8940006124138797, "learning_rate": 3.0977433725620633e-06, "loss": 1.1092, "step": 2872 }, { "epoch": 1.924313462826524, "grad_norm": 2.9296606919311685, "learning_rate": 3.0963241820007273e-06, "loss": 1.1829, "step": 2873 }, { "epoch": 1.9249832551908908, "grad_norm": 4.221458934603359, "learning_rate": 3.094904787647533e-06, "loss": 1.4622, "step": 2874 }, { "epoch": 1.925653047555258, "grad_norm": 3.850026662520149, "learning_rate": 3.093485189987554e-06, "loss": 1.3673, "step": 2875 }, { "epoch": 1.9263228399196248, "grad_norm": 2.2092597165285235, "learning_rate": 3.092065389505933e-06, "loss": 1.4148, "step": 2876 }, { "epoch": 1.926992632283992, "grad_norm": 2.165040997313289, "learning_rate": 3.090645386687881e-06, "loss": 1.1469, "step": 2877 }, { "epoch": 1.927662424648359, "grad_norm": 3.4870243090719684, "learning_rate": 3.0892251820186814e-06, "loss": 1.0029, "step": 2878 }, { "epoch": 1.9283322170127262, "grad_norm": 3.1270109722715995, "learning_rate": 3.087804775983683e-06, "loss": 1.2314, "step": 2879 }, { "epoch": 1.929002009377093, "grad_norm": 2.2997008351163477, "learning_rate": 3.0863841690683054e-06, "loss": 1.2541, "step": 2880 }, { "epoch": 1.9296718017414602, "grad_norm": 2.3074248399600754, "learning_rate": 3.084963361758037e-06, "loss": 1.2175, "step": 2881 }, { "epoch": 1.930341594105827, "grad_norm": 2.3317212555672215, "learning_rate": 3.0835423545384333e-06, "loss": 1.266, "step": 2882 }, { "epoch": 1.9310113864701943, "grad_norm": 3.82549177376904, "learning_rate": 3.0821211478951197e-06, "loss": 1.2344, "step": 2883 }, { "epoch": 1.9316811788345611, "grad_norm": 2.4172987522091796, "learning_rate": 3.0806997423137886e-06, "loss": 1.234, "step": 2884 }, { "epoch": 1.9323509711989284, "grad_norm": 7.768266514331327, "learning_rate": 3.0792781382802e-06, "loss": 1.1964, "step": 2885 }, { "epoch": 1.9330207635632952, "grad_norm": 4.178630692877541, "learning_rate": 3.0778563362801835e-06, "loss": 1.4517, "step": 2886 }, { "epoch": 1.9336905559276625, "grad_norm": 2.970084432614771, "learning_rate": 3.0764343367996346e-06, "loss": 1.2354, "step": 2887 }, { "epoch": 1.9343603482920295, "grad_norm": 2.3058089380134255, "learning_rate": 3.075012140324518e-06, "loss": 1.2776, "step": 2888 }, { "epoch": 1.9350301406563966, "grad_norm": 2.089663103937679, "learning_rate": 3.0735897473408637e-06, "loss": 1.3175, "step": 2889 }, { "epoch": 1.9356999330207636, "grad_norm": 2.261462177476998, "learning_rate": 3.072167158334771e-06, "loss": 1.2559, "step": 2890 }, { "epoch": 1.9363697253851306, "grad_norm": 3.578232860694065, "learning_rate": 3.070744373792403e-06, "loss": 1.153, "step": 2891 }, { "epoch": 1.9370395177494977, "grad_norm": 2.188770905905152, "learning_rate": 3.0693213941999945e-06, "loss": 1.3035, "step": 2892 }, { "epoch": 1.9377093101138647, "grad_norm": 3.4363317542326772, "learning_rate": 3.0678982200438418e-06, "loss": 1.3086, "step": 2893 }, { "epoch": 1.9383791024782318, "grad_norm": 2.8856078525918916, "learning_rate": 3.066474851810312e-06, "loss": 1.2522, "step": 2894 }, { "epoch": 1.9390488948425988, "grad_norm": 2.890217366559061, "learning_rate": 3.065051289985835e-06, "loss": 1.3965, "step": 2895 }, { "epoch": 1.9397186872069658, "grad_norm": 2.285103329101551, "learning_rate": 3.0636275350569094e-06, "loss": 1.2939, "step": 2896 }, { "epoch": 1.9403884795713329, "grad_norm": 2.042411493325065, "learning_rate": 3.062203587510098e-06, "loss": 1.2018, "step": 2897 }, { "epoch": 1.9410582719357, "grad_norm": 2.214500204556098, "learning_rate": 3.0607794478320316e-06, "loss": 1.2831, "step": 2898 }, { "epoch": 1.941728064300067, "grad_norm": 2.0879560632088747, "learning_rate": 3.0593551165094036e-06, "loss": 1.2414, "step": 2899 }, { "epoch": 1.942397856664434, "grad_norm": 3.406900795995699, "learning_rate": 3.0579305940289768e-06, "loss": 1.2438, "step": 2900 }, { "epoch": 1.943067649028801, "grad_norm": 8.435583335700432, "learning_rate": 3.0565058808775744e-06, "loss": 1.1822, "step": 2901 }, { "epoch": 1.943737441393168, "grad_norm": 2.724688609249763, "learning_rate": 3.05508097754209e-06, "loss": 0.9979, "step": 2902 }, { "epoch": 1.944407233757535, "grad_norm": 2.62416204697238, "learning_rate": 3.053655884509478e-06, "loss": 1.2778, "step": 2903 }, { "epoch": 1.9450770261219024, "grad_norm": 2.432619415448015, "learning_rate": 3.052230602266761e-06, "loss": 1.2577, "step": 2904 }, { "epoch": 1.9457468184862692, "grad_norm": 3.829707353479345, "learning_rate": 3.050805131301023e-06, "loss": 1.3278, "step": 2905 }, { "epoch": 1.9464166108506364, "grad_norm": 4.488631961484261, "learning_rate": 3.049379472099414e-06, "loss": 1.1827, "step": 2906 }, { "epoch": 1.9470864032150033, "grad_norm": 2.2315894616038014, "learning_rate": 3.0479536251491503e-06, "loss": 1.3206, "step": 2907 }, { "epoch": 1.9477561955793705, "grad_norm": 2.109345911929139, "learning_rate": 3.0465275909375087e-06, "loss": 1.323, "step": 2908 }, { "epoch": 1.9484259879437373, "grad_norm": 2.8759115110754414, "learning_rate": 3.045101369951832e-06, "loss": 1.0436, "step": 2909 }, { "epoch": 1.9490957803081046, "grad_norm": 2.9361785386234387, "learning_rate": 3.043674962679528e-06, "loss": 1.1534, "step": 2910 }, { "epoch": 1.9497655726724714, "grad_norm": 6.262153784204064, "learning_rate": 3.042248369608065e-06, "loss": 1.144, "step": 2911 }, { "epoch": 1.9504353650368387, "grad_norm": 2.4099096305468546, "learning_rate": 3.0408215912249777e-06, "loss": 1.2977, "step": 2912 }, { "epoch": 1.9511051574012055, "grad_norm": 2.5896679321373957, "learning_rate": 3.0393946280178623e-06, "loss": 1.2988, "step": 2913 }, { "epoch": 1.9517749497655728, "grad_norm": 2.1779184676119385, "learning_rate": 3.0379674804743793e-06, "loss": 1.4231, "step": 2914 }, { "epoch": 1.9524447421299396, "grad_norm": 2.0328729076988274, "learning_rate": 3.036540149082251e-06, "loss": 1.1646, "step": 2915 }, { "epoch": 1.9531145344943068, "grad_norm": 2.3436422974190614, "learning_rate": 3.035112634329265e-06, "loss": 1.3379, "step": 2916 }, { "epoch": 1.9537843268586736, "grad_norm": 1.9446172959716146, "learning_rate": 3.033684936703268e-06, "loss": 1.1425, "step": 2917 }, { "epoch": 1.954454119223041, "grad_norm": 2.611938462221784, "learning_rate": 3.0322570566921714e-06, "loss": 1.2063, "step": 2918 }, { "epoch": 1.955123911587408, "grad_norm": 2.8908777076436922, "learning_rate": 3.03082899478395e-06, "loss": 1.1963, "step": 2919 }, { "epoch": 1.955793703951775, "grad_norm": 3.068723244475811, "learning_rate": 3.0294007514666373e-06, "loss": 1.2212, "step": 2920 }, { "epoch": 1.956463496316142, "grad_norm": 2.993878293967151, "learning_rate": 3.0279723272283323e-06, "loss": 1.3484, "step": 2921 }, { "epoch": 1.957133288680509, "grad_norm": 5.211232594234047, "learning_rate": 3.0265437225571937e-06, "loss": 1.2815, "step": 2922 }, { "epoch": 1.957803081044876, "grad_norm": 2.4371211703905162, "learning_rate": 3.025114937941443e-06, "loss": 1.2701, "step": 2923 }, { "epoch": 1.9584728734092431, "grad_norm": 3.736201681025656, "learning_rate": 3.023685973869362e-06, "loss": 1.1535, "step": 2924 }, { "epoch": 1.9591426657736102, "grad_norm": 2.072748227744407, "learning_rate": 3.022256830829295e-06, "loss": 1.3827, "step": 2925 }, { "epoch": 1.9598124581379772, "grad_norm": 3.3546509944096217, "learning_rate": 3.0208275093096474e-06, "loss": 1.1469, "step": 2926 }, { "epoch": 1.9604822505023443, "grad_norm": 2.3267607423529677, "learning_rate": 3.019398009798884e-06, "loss": 1.2468, "step": 2927 }, { "epoch": 1.9611520428667113, "grad_norm": 4.214986822922903, "learning_rate": 3.0179683327855325e-06, "loss": 1.1895, "step": 2928 }, { "epoch": 1.9618218352310783, "grad_norm": 2.9423740608426274, "learning_rate": 3.0165384787581807e-06, "loss": 1.1684, "step": 2929 }, { "epoch": 1.9624916275954454, "grad_norm": 2.7170701104879034, "learning_rate": 3.0151084482054752e-06, "loss": 1.0884, "step": 2930 }, { "epoch": 1.9631614199598124, "grad_norm": 2.4799903820646776, "learning_rate": 3.013678241616127e-06, "loss": 1.3989, "step": 2931 }, { "epoch": 1.9638312123241795, "grad_norm": 3.592407923074649, "learning_rate": 3.012247859478901e-06, "loss": 1.069, "step": 2932 }, { "epoch": 1.9645010046885467, "grad_norm": 2.332799530017718, "learning_rate": 3.0108173022826294e-06, "loss": 1.3654, "step": 2933 }, { "epoch": 1.9651707970529135, "grad_norm": 2.227460933483187, "learning_rate": 3.009386570516197e-06, "loss": 1.2042, "step": 2934 }, { "epoch": 1.9658405894172808, "grad_norm": 9.34786215226198, "learning_rate": 3.007955664668554e-06, "loss": 1.3721, "step": 2935 }, { "epoch": 1.9665103817816476, "grad_norm": 4.488594629683764, "learning_rate": 3.0065245852287066e-06, "loss": 1.1687, "step": 2936 }, { "epoch": 1.9671801741460149, "grad_norm": 3.171236733987519, "learning_rate": 3.0050933326857233e-06, "loss": 1.3844, "step": 2937 }, { "epoch": 1.9678499665103817, "grad_norm": 2.6010299364779033, "learning_rate": 3.003661907528728e-06, "loss": 1.1831, "step": 2938 }, { "epoch": 1.968519758874749, "grad_norm": 2.3580610893143907, "learning_rate": 3.0022303102469074e-06, "loss": 1.3012, "step": 2939 }, { "epoch": 1.9691895512391158, "grad_norm": 4.917842396721976, "learning_rate": 3.000798541329503e-06, "loss": 1.2481, "step": 2940 }, { "epoch": 1.969859343603483, "grad_norm": 3.632905128016861, "learning_rate": 2.9993666012658196e-06, "loss": 1.1097, "step": 2941 }, { "epoch": 1.9705291359678498, "grad_norm": 2.3239719737310702, "learning_rate": 2.9979344905452157e-06, "loss": 1.2769, "step": 2942 }, { "epoch": 1.971198928332217, "grad_norm": 2.1645689849143412, "learning_rate": 2.9965022096571132e-06, "loss": 1.1995, "step": 2943 }, { "epoch": 1.971868720696584, "grad_norm": 2.2683367974891757, "learning_rate": 2.995069759090987e-06, "loss": 1.1816, "step": 2944 }, { "epoch": 1.9725385130609512, "grad_norm": 2.1375549604967103, "learning_rate": 2.993637139336375e-06, "loss": 1.2173, "step": 2945 }, { "epoch": 1.973208305425318, "grad_norm": 4.496444591383753, "learning_rate": 2.992204350882867e-06, "loss": 1.3311, "step": 2946 }, { "epoch": 1.9738780977896853, "grad_norm": 2.159941950718647, "learning_rate": 2.990771394220117e-06, "loss": 1.1945, "step": 2947 }, { "epoch": 1.9745478901540523, "grad_norm": 3.0175140388553876, "learning_rate": 2.989338269837831e-06, "loss": 1.2521, "step": 2948 }, { "epoch": 1.9752176825184193, "grad_norm": 2.402790071472187, "learning_rate": 2.9879049782257765e-06, "loss": 1.4113, "step": 2949 }, { "epoch": 1.9758874748827864, "grad_norm": 2.581182022602077, "learning_rate": 2.986471519873775e-06, "loss": 1.1533, "step": 2950 }, { "epoch": 1.9765572672471534, "grad_norm": 3.251238644078216, "learning_rate": 2.9850378952717063e-06, "loss": 1.3408, "step": 2951 }, { "epoch": 1.9772270596115205, "grad_norm": 2.3281448328011187, "learning_rate": 2.9836041049095073e-06, "loss": 1.2973, "step": 2952 }, { "epoch": 1.9778968519758875, "grad_norm": 2.3626269042414134, "learning_rate": 2.9821701492771717e-06, "loss": 1.2684, "step": 2953 }, { "epoch": 1.9785666443402545, "grad_norm": 2.2260983671474235, "learning_rate": 2.980736028864748e-06, "loss": 1.3323, "step": 2954 }, { "epoch": 1.9792364367046216, "grad_norm": 2.821956269429714, "learning_rate": 2.9793017441623436e-06, "loss": 1.1417, "step": 2955 }, { "epoch": 1.9799062290689886, "grad_norm": 3.148584266976301, "learning_rate": 2.9778672956601194e-06, "loss": 1.1274, "step": 2956 }, { "epoch": 1.9805760214333556, "grad_norm": 2.2605100569783256, "learning_rate": 2.9764326838482943e-06, "loss": 1.0977, "step": 2957 }, { "epoch": 1.9812458137977227, "grad_norm": 16.345361061096238, "learning_rate": 2.9749979092171422e-06, "loss": 1.2013, "step": 2958 }, { "epoch": 1.9819156061620897, "grad_norm": 3.371520139292595, "learning_rate": 2.973562972256992e-06, "loss": 1.3242, "step": 2959 }, { "epoch": 1.9825853985264568, "grad_norm": 2.4584522588488475, "learning_rate": 2.9721278734582297e-06, "loss": 1.4856, "step": 2960 }, { "epoch": 1.9832551908908238, "grad_norm": 2.3671839829544115, "learning_rate": 2.970692613311295e-06, "loss": 1.1419, "step": 2961 }, { "epoch": 1.9839249832551908, "grad_norm": 3.6093766293868064, "learning_rate": 2.9692571923066838e-06, "loss": 1.2142, "step": 2962 }, { "epoch": 1.9845947756195579, "grad_norm": 2.577607691534303, "learning_rate": 2.9678216109349465e-06, "loss": 1.0986, "step": 2963 }, { "epoch": 1.9852645679839251, "grad_norm": 6.1588962798751385, "learning_rate": 2.9663858696866883e-06, "loss": 1.0333, "step": 2964 }, { "epoch": 1.985934360348292, "grad_norm": 2.135975931953536, "learning_rate": 2.9649499690525686e-06, "loss": 1.355, "step": 2965 }, { "epoch": 1.9866041527126592, "grad_norm": 2.3290349030480098, "learning_rate": 2.963513909523303e-06, "loss": 1.2099, "step": 2966 }, { "epoch": 1.987273945077026, "grad_norm": 2.2645692508007085, "learning_rate": 2.9620776915896587e-06, "loss": 1.2421, "step": 2967 }, { "epoch": 1.9879437374413933, "grad_norm": 3.2859920891221015, "learning_rate": 2.9606413157424605e-06, "loss": 1.2574, "step": 2968 }, { "epoch": 1.9886135298057601, "grad_norm": 2.138152753766078, "learning_rate": 2.9592047824725827e-06, "loss": 1.3187, "step": 2969 }, { "epoch": 1.9892833221701274, "grad_norm": 2.250988229236177, "learning_rate": 2.9577680922709584e-06, "loss": 1.2726, "step": 2970 }, { "epoch": 1.9899531145344942, "grad_norm": 2.147862086858305, "learning_rate": 2.9563312456285697e-06, "loss": 1.2156, "step": 2971 }, { "epoch": 1.9906229068988615, "grad_norm": 2.531933518824578, "learning_rate": 2.954894243036457e-06, "loss": 1.1106, "step": 2972 }, { "epoch": 1.9912926992632283, "grad_norm": 3.177248363794571, "learning_rate": 2.9534570849857074e-06, "loss": 1.1749, "step": 2973 }, { "epoch": 1.9919624916275955, "grad_norm": 2.827075340059879, "learning_rate": 2.952019771967469e-06, "loss": 1.3443, "step": 2974 }, { "epoch": 1.9926322839919623, "grad_norm": 2.524242025414766, "learning_rate": 2.9505823044729353e-06, "loss": 1.1011, "step": 2975 }, { "epoch": 1.9933020763563296, "grad_norm": 2.480262018563047, "learning_rate": 2.949144682993359e-06, "loss": 1.401, "step": 2976 }, { "epoch": 1.9939718687206964, "grad_norm": 3.850564269651228, "learning_rate": 2.9477069080200406e-06, "loss": 1.3725, "step": 2977 }, { "epoch": 1.9946416610850637, "grad_norm": 2.180208897405579, "learning_rate": 2.946268980044337e-06, "loss": 1.252, "step": 2978 }, { "epoch": 1.9953114534494307, "grad_norm": 2.207084639895355, "learning_rate": 2.944830899557653e-06, "loss": 1.1394, "step": 2979 }, { "epoch": 1.9959812458137978, "grad_norm": 2.0778053471971623, "learning_rate": 2.9433926670514507e-06, "loss": 1.3343, "step": 2980 }, { "epoch": 1.9966510381781648, "grad_norm": 3.3532175826123005, "learning_rate": 2.9419542830172386e-06, "loss": 1.2701, "step": 2981 }, { "epoch": 1.9973208305425318, "grad_norm": 2.397444213636326, "learning_rate": 2.9405157479465823e-06, "loss": 1.3064, "step": 2982 }, { "epoch": 1.9979906229068989, "grad_norm": 1.9862798282221061, "learning_rate": 2.939077062331095e-06, "loss": 1.2458, "step": 2983 }, { "epoch": 1.998660415271266, "grad_norm": 2.2593490930615716, "learning_rate": 2.9376382266624437e-06, "loss": 1.4908, "step": 2984 }, { "epoch": 1.999330207635633, "grad_norm": 2.7125501927909097, "learning_rate": 2.9361992414323443e-06, "loss": 1.1766, "step": 2985 }, { "epoch": 2.0, "grad_norm": 2.6402906810405216, "learning_rate": 2.934760107132567e-06, "loss": 1.3163, "step": 2986 }, { "epoch": 2.0006697923643673, "grad_norm": 2.1370119075777674, "learning_rate": 2.9333208242549303e-06, "loss": 1.2353, "step": 2987 }, { "epoch": 2.001339584728734, "grad_norm": 2.617209824771479, "learning_rate": 2.931881393291306e-06, "loss": 1.0599, "step": 2988 }, { "epoch": 2.0020093770931013, "grad_norm": 2.4289226412973304, "learning_rate": 2.930441814733612e-06, "loss": 1.0875, "step": 2989 }, { "epoch": 2.002679169457468, "grad_norm": 2.4783513117636433, "learning_rate": 2.929002089073823e-06, "loss": 1.1142, "step": 2990 }, { "epoch": 2.0033489618218354, "grad_norm": 2.3729082608839045, "learning_rate": 2.9275622168039582e-06, "loss": 1.2552, "step": 2991 }, { "epoch": 2.0040187541862022, "grad_norm": 2.44670529694116, "learning_rate": 2.926122198416091e-06, "loss": 1.2103, "step": 2992 }, { "epoch": 2.0046885465505695, "grad_norm": 3.0759898804499044, "learning_rate": 2.9246820344023406e-06, "loss": 1.2431, "step": 2993 }, { "epoch": 2.0053583389149363, "grad_norm": 2.675191308990021, "learning_rate": 2.923241725254881e-06, "loss": 1.1152, "step": 2994 }, { "epoch": 2.0060281312793036, "grad_norm": 2.210193137213327, "learning_rate": 2.9218012714659323e-06, "loss": 1.219, "step": 2995 }, { "epoch": 2.0066979236436704, "grad_norm": 2.608497684047579, "learning_rate": 2.9203606735277638e-06, "loss": 1.2203, "step": 2996 }, { "epoch": 2.0073677160080377, "grad_norm": 2.9945317615374845, "learning_rate": 2.9189199319326965e-06, "loss": 0.9957, "step": 2997 }, { "epoch": 2.0080375083724045, "grad_norm": 2.343593930995482, "learning_rate": 2.917479047173099e-06, "loss": 1.1837, "step": 2998 }, { "epoch": 2.0087073007367717, "grad_norm": 1.9073560532533749, "learning_rate": 2.916038019741389e-06, "loss": 0.8022, "step": 2999 }, { "epoch": 2.0093770931011385, "grad_norm": 2.623113018793343, "learning_rate": 2.9145968501300325e-06, "loss": 1.1739, "step": 3000 }, { "epoch": 2.010046885465506, "grad_norm": 2.4803599830276504, "learning_rate": 2.9131555388315447e-06, "loss": 1.1296, "step": 3001 }, { "epoch": 2.0107166778298726, "grad_norm": 2.389809459405461, "learning_rate": 2.9117140863384893e-06, "loss": 1.1342, "step": 3002 }, { "epoch": 2.01138647019424, "grad_norm": 2.1924096652162213, "learning_rate": 2.910272493143478e-06, "loss": 1.2208, "step": 3003 }, { "epoch": 2.0120562625586067, "grad_norm": 2.185391416600572, "learning_rate": 2.9088307597391695e-06, "loss": 1.0611, "step": 3004 }, { "epoch": 2.012726054922974, "grad_norm": 2.3621478169802352, "learning_rate": 2.9073888866182735e-06, "loss": 1.1279, "step": 3005 }, { "epoch": 2.013395847287341, "grad_norm": 2.433367412892989, "learning_rate": 2.905946874273544e-06, "loss": 1.1378, "step": 3006 }, { "epoch": 2.014065639651708, "grad_norm": 2.2098498133991042, "learning_rate": 2.9045047231977853e-06, "loss": 0.9519, "step": 3007 }, { "epoch": 2.014735432016075, "grad_norm": 2.106807489441195, "learning_rate": 2.9030624338838465e-06, "loss": 1.216, "step": 3008 }, { "epoch": 2.015405224380442, "grad_norm": 2.5858536511384216, "learning_rate": 2.9016200068246258e-06, "loss": 1.1812, "step": 3009 }, { "epoch": 2.016075016744809, "grad_norm": 2.381531639076817, "learning_rate": 2.900177442513068e-06, "loss": 1.0286, "step": 3010 }, { "epoch": 2.016744809109176, "grad_norm": 2.8171626846371582, "learning_rate": 2.8987347414421656e-06, "loss": 1.2857, "step": 3011 }, { "epoch": 2.017414601473543, "grad_norm": 2.084201549323835, "learning_rate": 2.897291904104955e-06, "loss": 0.9126, "step": 3012 }, { "epoch": 2.0180843938379103, "grad_norm": 2.325210219468293, "learning_rate": 2.8958489309945235e-06, "loss": 1.0927, "step": 3013 }, { "epoch": 2.018754186202277, "grad_norm": 2.2198795354083614, "learning_rate": 2.8944058226040013e-06, "loss": 1.1923, "step": 3014 }, { "epoch": 2.0194239785666444, "grad_norm": 2.418068295559066, "learning_rate": 2.8929625794265666e-06, "loss": 1.247, "step": 3015 }, { "epoch": 2.020093770931011, "grad_norm": 2.223291869806444, "learning_rate": 2.8915192019554416e-06, "loss": 0.9331, "step": 3016 }, { "epoch": 2.0207635632953784, "grad_norm": 2.5739040677146887, "learning_rate": 2.890075690683898e-06, "loss": 1.0263, "step": 3017 }, { "epoch": 2.0214333556597457, "grad_norm": 3.463985339840248, "learning_rate": 2.8886320461052487e-06, "loss": 1.0486, "step": 3018 }, { "epoch": 2.0221031480241125, "grad_norm": 2.721619527159129, "learning_rate": 2.8871882687128573e-06, "loss": 1.16, "step": 3019 }, { "epoch": 2.0227729403884798, "grad_norm": 3.3267833656966364, "learning_rate": 2.8857443590001277e-06, "loss": 0.8168, "step": 3020 }, { "epoch": 2.0234427327528466, "grad_norm": 2.503104931270395, "learning_rate": 2.8843003174605126e-06, "loss": 1.2005, "step": 3021 }, { "epoch": 2.024112525117214, "grad_norm": 2.5532517625070157, "learning_rate": 2.882856144587508e-06, "loss": 1.1063, "step": 3022 }, { "epoch": 2.0247823174815807, "grad_norm": 2.1230771392680223, "learning_rate": 2.8814118408746566e-06, "loss": 1.1022, "step": 3023 }, { "epoch": 2.025452109845948, "grad_norm": 2.3977839511211, "learning_rate": 2.879967406815542e-06, "loss": 1.2498, "step": 3024 }, { "epoch": 2.0261219022103147, "grad_norm": 7.362649473750122, "learning_rate": 2.8785228429037974e-06, "loss": 1.1306, "step": 3025 }, { "epoch": 2.026791694574682, "grad_norm": 2.5835179988298527, "learning_rate": 2.8770781496330963e-06, "loss": 1.0381, "step": 3026 }, { "epoch": 2.027461486939049, "grad_norm": 3.9404119447079333, "learning_rate": 2.8756333274971587e-06, "loss": 0.9495, "step": 3027 }, { "epoch": 2.028131279303416, "grad_norm": 2.0785532293575035, "learning_rate": 2.874188376989747e-06, "loss": 1.0698, "step": 3028 }, { "epoch": 2.028801071667783, "grad_norm": 2.2792320122570677, "learning_rate": 2.87274329860467e-06, "loss": 1.1926, "step": 3029 }, { "epoch": 2.02947086403215, "grad_norm": 1.948976164633069, "learning_rate": 2.8712980928357765e-06, "loss": 0.9388, "step": 3030 }, { "epoch": 2.030140656396517, "grad_norm": 2.4458490573054177, "learning_rate": 2.8698527601769625e-06, "loss": 1.0543, "step": 3031 }, { "epoch": 2.0308104487608842, "grad_norm": 2.7909114227857104, "learning_rate": 2.8684073011221646e-06, "loss": 1.1081, "step": 3032 }, { "epoch": 2.031480241125251, "grad_norm": 2.49918649254412, "learning_rate": 2.8669617161653653e-06, "loss": 1.0011, "step": 3033 }, { "epoch": 2.0321500334896183, "grad_norm": 2.0955702998979615, "learning_rate": 2.865516005800586e-06, "loss": 1.1712, "step": 3034 }, { "epoch": 2.032819825853985, "grad_norm": 2.5349701257851818, "learning_rate": 2.8640701705218966e-06, "loss": 1.2963, "step": 3035 }, { "epoch": 2.0334896182183524, "grad_norm": 4.227503435028214, "learning_rate": 2.8626242108234044e-06, "loss": 0.9612, "step": 3036 }, { "epoch": 2.034159410582719, "grad_norm": 2.210391617503558, "learning_rate": 2.861178127199262e-06, "loss": 1.1832, "step": 3037 }, { "epoch": 2.0348292029470865, "grad_norm": 2.244312741799867, "learning_rate": 2.8597319201436645e-06, "loss": 1.1982, "step": 3038 }, { "epoch": 2.0354989953114533, "grad_norm": 2.931751186913981, "learning_rate": 2.8582855901508484e-06, "loss": 1.1147, "step": 3039 }, { "epoch": 2.0361687876758205, "grad_norm": 2.714572628686955, "learning_rate": 2.856839137715091e-06, "loss": 1.186, "step": 3040 }, { "epoch": 2.0368385800401874, "grad_norm": 2.163253804515489, "learning_rate": 2.8553925633307143e-06, "loss": 1.1573, "step": 3041 }, { "epoch": 2.0375083724045546, "grad_norm": 2.7389751308011463, "learning_rate": 2.8539458674920795e-06, "loss": 1.0474, "step": 3042 }, { "epoch": 2.0381781647689214, "grad_norm": 4.573407814012291, "learning_rate": 2.85249905069359e-06, "loss": 1.2487, "step": 3043 }, { "epoch": 2.0388479571332887, "grad_norm": 2.3343062806995993, "learning_rate": 2.851052113429692e-06, "loss": 1.2293, "step": 3044 }, { "epoch": 2.0395177494976555, "grad_norm": 3.2724818352458644, "learning_rate": 2.84960505619487e-06, "loss": 0.9971, "step": 3045 }, { "epoch": 2.040187541862023, "grad_norm": 2.894336942942175, "learning_rate": 2.8481578794836523e-06, "loss": 1.0748, "step": 3046 }, { "epoch": 2.0408573342263896, "grad_norm": 2.37922083980148, "learning_rate": 2.8467105837906067e-06, "loss": 1.1395, "step": 3047 }, { "epoch": 2.041527126590757, "grad_norm": 2.4216262223645084, "learning_rate": 2.8452631696103417e-06, "loss": 1.0861, "step": 3048 }, { "epoch": 2.042196918955124, "grad_norm": 2.3589715020570226, "learning_rate": 2.8438156374375057e-06, "loss": 1.1761, "step": 3049 }, { "epoch": 2.042866711319491, "grad_norm": 2.465592679810069, "learning_rate": 2.8423679877667894e-06, "loss": 1.2402, "step": 3050 }, { "epoch": 2.043536503683858, "grad_norm": 5.662877928126193, "learning_rate": 2.840920221092921e-06, "loss": 0.9771, "step": 3051 }, { "epoch": 2.044206296048225, "grad_norm": 2.2728343393362183, "learning_rate": 2.8394723379106714e-06, "loss": 0.9949, "step": 3052 }, { "epoch": 2.0448760884125923, "grad_norm": 3.1419064587850802, "learning_rate": 2.8380243387148494e-06, "loss": 1.0974, "step": 3053 }, { "epoch": 2.045545880776959, "grad_norm": 2.2088442366395507, "learning_rate": 2.8365762240003043e-06, "loss": 1.1524, "step": 3054 }, { "epoch": 2.0462156731413264, "grad_norm": 2.494515472055137, "learning_rate": 2.8351279942619232e-06, "loss": 1.1692, "step": 3055 }, { "epoch": 2.046885465505693, "grad_norm": 2.203560130639722, "learning_rate": 2.8336796499946363e-06, "loss": 1.1566, "step": 3056 }, { "epoch": 2.0475552578700604, "grad_norm": 2.5459213384023585, "learning_rate": 2.8322311916934086e-06, "loss": 1.3257, "step": 3057 }, { "epoch": 2.0482250502344272, "grad_norm": 3.0122780175562056, "learning_rate": 2.8307826198532474e-06, "loss": 1.1895, "step": 3058 }, { "epoch": 2.0488948425987945, "grad_norm": 2.0513805062478725, "learning_rate": 2.829333934969196e-06, "loss": 1.0314, "step": 3059 }, { "epoch": 2.0495646349631613, "grad_norm": 2.1628401925306022, "learning_rate": 2.82788513753634e-06, "loss": 1.1272, "step": 3060 }, { "epoch": 2.0502344273275286, "grad_norm": 2.2586754011652777, "learning_rate": 2.826436228049798e-06, "loss": 1.0504, "step": 3061 }, { "epoch": 2.0509042196918954, "grad_norm": 4.564806908468202, "learning_rate": 2.824987207004734e-06, "loss": 1.1026, "step": 3062 }, { "epoch": 2.0515740120562627, "grad_norm": 2.2680119827481002, "learning_rate": 2.8235380748963425e-06, "loss": 1.3335, "step": 3063 }, { "epoch": 2.0522438044206295, "grad_norm": 2.0805665712222923, "learning_rate": 2.8220888322198624e-06, "loss": 1.0708, "step": 3064 }, { "epoch": 2.0529135967849967, "grad_norm": 2.381550861978096, "learning_rate": 2.820639479470566e-06, "loss": 1.1404, "step": 3065 }, { "epoch": 2.0535833891493636, "grad_norm": 2.1276129317747023, "learning_rate": 2.819190017143766e-06, "loss": 1.1898, "step": 3066 }, { "epoch": 2.054253181513731, "grad_norm": 3.5541084679841273, "learning_rate": 2.817740445734811e-06, "loss": 1.0307, "step": 3067 }, { "epoch": 2.0549229738780976, "grad_norm": 2.445671342013536, "learning_rate": 2.8162907657390882e-06, "loss": 1.2353, "step": 3068 }, { "epoch": 2.055592766242465, "grad_norm": 2.209632160498867, "learning_rate": 2.8148409776520185e-06, "loss": 1.322, "step": 3069 }, { "epoch": 2.0562625586068317, "grad_norm": 4.03620778415916, "learning_rate": 2.8133910819690656e-06, "loss": 1.0293, "step": 3070 }, { "epoch": 2.056932350971199, "grad_norm": 2.126959656656598, "learning_rate": 2.8119410791857244e-06, "loss": 0.9896, "step": 3071 }, { "epoch": 2.057602143335566, "grad_norm": 2.251906410785338, "learning_rate": 2.810490969797529e-06, "loss": 1.1741, "step": 3072 }, { "epoch": 2.058271935699933, "grad_norm": 2.211831224973581, "learning_rate": 2.8090407543000504e-06, "loss": 1.1838, "step": 3073 }, { "epoch": 2.0589417280643, "grad_norm": 2.1896856235575926, "learning_rate": 2.8075904331888946e-06, "loss": 1.1985, "step": 3074 }, { "epoch": 2.059611520428667, "grad_norm": 2.2146089182142203, "learning_rate": 2.8061400069597046e-06, "loss": 1.266, "step": 3075 }, { "epoch": 2.060281312793034, "grad_norm": 2.2796021618989557, "learning_rate": 2.804689476108157e-06, "loss": 1.0386, "step": 3076 }, { "epoch": 2.060951105157401, "grad_norm": 4.046137703005577, "learning_rate": 2.8032388411299686e-06, "loss": 1.2418, "step": 3077 }, { "epoch": 2.0616208975217685, "grad_norm": 3.414665495112858, "learning_rate": 2.8017881025208883e-06, "loss": 0.9015, "step": 3078 }, { "epoch": 2.0622906898861353, "grad_norm": 3.0655242903490048, "learning_rate": 2.8003372607766997e-06, "loss": 1.1524, "step": 3079 }, { "epoch": 2.0629604822505025, "grad_norm": 2.1983851844574556, "learning_rate": 2.7988863163932256e-06, "loss": 1.0881, "step": 3080 }, { "epoch": 2.0636302746148694, "grad_norm": 2.551247834344094, "learning_rate": 2.7974352698663203e-06, "loss": 1.1747, "step": 3081 }, { "epoch": 2.0643000669792366, "grad_norm": 2.4850174663872076, "learning_rate": 2.795984121691875e-06, "loss": 1.1639, "step": 3082 }, { "epoch": 2.0649698593436034, "grad_norm": 3.527503904103635, "learning_rate": 2.794532872365814e-06, "loss": 1.1142, "step": 3083 }, { "epoch": 2.0656396517079707, "grad_norm": 2.034602970544174, "learning_rate": 2.7930815223840973e-06, "loss": 1.0068, "step": 3084 }, { "epoch": 2.0663094440723375, "grad_norm": 2.153360070438238, "learning_rate": 2.7916300722427193e-06, "loss": 1.1108, "step": 3085 }, { "epoch": 2.066979236436705, "grad_norm": 2.3244122253121127, "learning_rate": 2.7901785224377083e-06, "loss": 1.2541, "step": 3086 }, { "epoch": 2.0676490288010716, "grad_norm": 2.1978548918243694, "learning_rate": 2.788726873465127e-06, "loss": 1.3066, "step": 3087 }, { "epoch": 2.068318821165439, "grad_norm": 2.222020844017803, "learning_rate": 2.7872751258210712e-06, "loss": 1.1051, "step": 3088 }, { "epoch": 2.0689886135298057, "grad_norm": 2.3002396958227607, "learning_rate": 2.7858232800016714e-06, "loss": 1.122, "step": 3089 }, { "epoch": 2.069658405894173, "grad_norm": 2.1809369602565796, "learning_rate": 2.784371336503091e-06, "loss": 1.1304, "step": 3090 }, { "epoch": 2.0703281982585398, "grad_norm": 2.3699921333661687, "learning_rate": 2.7829192958215274e-06, "loss": 1.0914, "step": 3091 }, { "epoch": 2.070997990622907, "grad_norm": 2.751194981815927, "learning_rate": 2.78146715845321e-06, "loss": 1.1463, "step": 3092 }, { "epoch": 2.071667782987274, "grad_norm": 2.602729827148616, "learning_rate": 2.7800149248944024e-06, "loss": 1.1085, "step": 3093 }, { "epoch": 2.072337575351641, "grad_norm": 2.5938673019464655, "learning_rate": 2.7785625956414008e-06, "loss": 1.2546, "step": 3094 }, { "epoch": 2.073007367716008, "grad_norm": 7.328416303335982, "learning_rate": 2.7771101711905348e-06, "loss": 1.1032, "step": 3095 }, { "epoch": 2.073677160080375, "grad_norm": 3.6022466817480856, "learning_rate": 2.775657652038164e-06, "loss": 1.0518, "step": 3096 }, { "epoch": 2.074346952444742, "grad_norm": 2.7097644690929603, "learning_rate": 2.7742050386806836e-06, "loss": 1.157, "step": 3097 }, { "epoch": 2.0750167448091092, "grad_norm": 2.3171939625268045, "learning_rate": 2.772752331614519e-06, "loss": 1.1712, "step": 3098 }, { "epoch": 2.075686537173476, "grad_norm": 3.3485346716557456, "learning_rate": 2.771299531336128e-06, "loss": 1.1046, "step": 3099 }, { "epoch": 2.0763563295378433, "grad_norm": 2.3680306266795963, "learning_rate": 2.7698466383420003e-06, "loss": 0.9461, "step": 3100 }, { "epoch": 2.07702612190221, "grad_norm": 2.2728908538346446, "learning_rate": 2.768393653128658e-06, "loss": 1.1684, "step": 3101 }, { "epoch": 2.0776959142665774, "grad_norm": 2.5482113579371957, "learning_rate": 2.7669405761926534e-06, "loss": 0.9897, "step": 3102 }, { "epoch": 2.078365706630944, "grad_norm": 2.892301162040708, "learning_rate": 2.765487408030572e-06, "loss": 1.2708, "step": 3103 }, { "epoch": 2.0790354989953115, "grad_norm": 2.981128692595722, "learning_rate": 2.764034149139027e-06, "loss": 1.02, "step": 3104 }, { "epoch": 2.0797052913596783, "grad_norm": 2.458570103628293, "learning_rate": 2.762580800014668e-06, "loss": 1.2953, "step": 3105 }, { "epoch": 2.0803750837240456, "grad_norm": 2.441877831212849, "learning_rate": 2.7611273611541695e-06, "loss": 1.0439, "step": 3106 }, { "epoch": 2.081044876088413, "grad_norm": 2.426722201383944, "learning_rate": 2.759673833054241e-06, "loss": 1.2649, "step": 3107 }, { "epoch": 2.0817146684527796, "grad_norm": 2.965303282181617, "learning_rate": 2.758220216211621e-06, "loss": 1.0599, "step": 3108 }, { "epoch": 2.082384460817147, "grad_norm": 2.239007229643003, "learning_rate": 2.7567665111230783e-06, "loss": 1.1796, "step": 3109 }, { "epoch": 2.0830542531815137, "grad_norm": 2.4598483276994663, "learning_rate": 2.755312718285412e-06, "loss": 1.0822, "step": 3110 }, { "epoch": 2.083724045545881, "grad_norm": 2.6822829324622424, "learning_rate": 2.75385883819545e-06, "loss": 1.1439, "step": 3111 }, { "epoch": 2.084393837910248, "grad_norm": 2.3573928693585193, "learning_rate": 2.752404871350052e-06, "loss": 1.0861, "step": 3112 }, { "epoch": 2.085063630274615, "grad_norm": 4.065271532779092, "learning_rate": 2.7509508182461076e-06, "loss": 1.1964, "step": 3113 }, { "epoch": 2.085733422638982, "grad_norm": 2.7403788853733637, "learning_rate": 2.749496679380532e-06, "loss": 1.1923, "step": 3114 }, { "epoch": 2.086403215003349, "grad_norm": 2.9206572545081046, "learning_rate": 2.7480424552502755e-06, "loss": 1.0989, "step": 3115 }, { "epoch": 2.087073007367716, "grad_norm": 2.205006785516216, "learning_rate": 2.746588146352312e-06, "loss": 1.1418, "step": 3116 }, { "epoch": 2.087742799732083, "grad_norm": 2.5145641084933437, "learning_rate": 2.745133753183648e-06, "loss": 1.1705, "step": 3117 }, { "epoch": 2.08841259209645, "grad_norm": 2.410000634941881, "learning_rate": 2.743679276241318e-06, "loss": 1.0641, "step": 3118 }, { "epoch": 2.0890823844608173, "grad_norm": 2.449898984636239, "learning_rate": 2.7422247160223838e-06, "loss": 0.8753, "step": 3119 }, { "epoch": 2.089752176825184, "grad_norm": 2.1620935147006386, "learning_rate": 2.7407700730239374e-06, "loss": 0.9664, "step": 3120 }, { "epoch": 2.0904219691895514, "grad_norm": 2.157635771510232, "learning_rate": 2.739315347743098e-06, "loss": 1.2527, "step": 3121 }, { "epoch": 2.091091761553918, "grad_norm": 2.761948515345515, "learning_rate": 2.737860540677013e-06, "loss": 0.9183, "step": 3122 }, { "epoch": 2.0917615539182854, "grad_norm": 2.334885406089793, "learning_rate": 2.736405652322859e-06, "loss": 1.2062, "step": 3123 }, { "epoch": 2.0924313462826523, "grad_norm": 1.9889631583546692, "learning_rate": 2.7349506831778375e-06, "loss": 0.9107, "step": 3124 }, { "epoch": 2.0931011386470195, "grad_norm": 2.2205700320046073, "learning_rate": 2.7334956337391823e-06, "loss": 1.3291, "step": 3125 }, { "epoch": 2.0937709310113863, "grad_norm": 3.0093454192636653, "learning_rate": 2.732040504504149e-06, "loss": 1.2468, "step": 3126 }, { "epoch": 2.0944407233757536, "grad_norm": 2.31072446074402, "learning_rate": 2.7305852959700254e-06, "loss": 1.0576, "step": 3127 }, { "epoch": 2.0951105157401204, "grad_norm": 2.9036053331591933, "learning_rate": 2.7291300086341226e-06, "loss": 0.9789, "step": 3128 }, { "epoch": 2.0957803081044877, "grad_norm": 2.358086938876988, "learning_rate": 2.7276746429937817e-06, "loss": 1.1002, "step": 3129 }, { "epoch": 2.0964501004688545, "grad_norm": 2.4971357951349185, "learning_rate": 2.726219199546369e-06, "loss": 0.9307, "step": 3130 }, { "epoch": 2.0971198928332218, "grad_norm": 2.364828187541424, "learning_rate": 2.7247636787892766e-06, "loss": 0.9701, "step": 3131 }, { "epoch": 2.0977896851975886, "grad_norm": 2.3387693457137173, "learning_rate": 2.723308081219925e-06, "loss": 1.1782, "step": 3132 }, { "epoch": 2.098459477561956, "grad_norm": 4.161954908504713, "learning_rate": 2.72185240733576e-06, "loss": 1.1581, "step": 3133 }, { "epoch": 2.0991292699263226, "grad_norm": 2.2190949545620366, "learning_rate": 2.7203966576342526e-06, "loss": 1.0552, "step": 3134 }, { "epoch": 2.09979906229069, "grad_norm": 3.013567619377278, "learning_rate": 2.7189408326129014e-06, "loss": 0.9938, "step": 3135 }, { "epoch": 2.1004688546550567, "grad_norm": 2.3648315155929587, "learning_rate": 2.717484932769229e-06, "loss": 1.1831, "step": 3136 }, { "epoch": 2.101138647019424, "grad_norm": 2.429254296895792, "learning_rate": 2.7160289586007854e-06, "loss": 1.1199, "step": 3137 }, { "epoch": 2.101808439383791, "grad_norm": 2.4499067690924403, "learning_rate": 2.7145729106051445e-06, "loss": 1.1091, "step": 3138 }, { "epoch": 2.102478231748158, "grad_norm": 2.513331277046493, "learning_rate": 2.713116789279906e-06, "loss": 1.1607, "step": 3139 }, { "epoch": 2.1031480241125253, "grad_norm": 3.6629508842683403, "learning_rate": 2.711660595122695e-06, "loss": 1.0629, "step": 3140 }, { "epoch": 2.103817816476892, "grad_norm": 2.1154790788776245, "learning_rate": 2.71020432863116e-06, "loss": 1.0561, "step": 3141 }, { "epoch": 2.1044876088412594, "grad_norm": 2.764208954055331, "learning_rate": 2.7087479903029767e-06, "loss": 1.1671, "step": 3142 }, { "epoch": 2.105157401205626, "grad_norm": 2.187735205582294, "learning_rate": 2.7072915806358446e-06, "loss": 1.1785, "step": 3143 }, { "epoch": 2.1058271935699935, "grad_norm": 2.267027670963344, "learning_rate": 2.7058351001274847e-06, "loss": 1.1271, "step": 3144 }, { "epoch": 2.1064969859343603, "grad_norm": 2.4451827282579766, "learning_rate": 2.7043785492756464e-06, "loss": 1.2294, "step": 3145 }, { "epoch": 2.1071667782987276, "grad_norm": 2.1235128825253047, "learning_rate": 2.7029219285781e-06, "loss": 1.1097, "step": 3146 }, { "epoch": 2.1078365706630944, "grad_norm": 2.436070648919979, "learning_rate": 2.7014652385326425e-06, "loss": 1.1028, "step": 3147 }, { "epoch": 2.1085063630274616, "grad_norm": 2.5670218300837875, "learning_rate": 2.700008479637091e-06, "loss": 1.2144, "step": 3148 }, { "epoch": 2.1091761553918285, "grad_norm": 2.6331235867577605, "learning_rate": 2.69855165238929e-06, "loss": 1.1185, "step": 3149 }, { "epoch": 2.1098459477561957, "grad_norm": 3.9562111809890346, "learning_rate": 2.6970947572871043e-06, "loss": 1.1972, "step": 3150 }, { "epoch": 2.1105157401205625, "grad_norm": 2.520654134316489, "learning_rate": 2.695637794828423e-06, "loss": 1.1351, "step": 3151 }, { "epoch": 2.11118553248493, "grad_norm": 2.1794999993112065, "learning_rate": 2.694180765511159e-06, "loss": 1.0287, "step": 3152 }, { "epoch": 2.1118553248492966, "grad_norm": 2.242798737636118, "learning_rate": 2.692723669833247e-06, "loss": 1.0803, "step": 3153 }, { "epoch": 2.112525117213664, "grad_norm": 2.4989560259179284, "learning_rate": 2.6912665082926452e-06, "loss": 1.0665, "step": 3154 }, { "epoch": 2.1131949095780307, "grad_norm": 2.3459314604409345, "learning_rate": 2.6898092813873335e-06, "loss": 1.1275, "step": 3155 }, { "epoch": 2.113864701942398, "grad_norm": 2.3410999251063007, "learning_rate": 2.6883519896153137e-06, "loss": 1.1523, "step": 3156 }, { "epoch": 2.1145344943067648, "grad_norm": 2.3413216071926755, "learning_rate": 2.686894633474612e-06, "loss": 1.1738, "step": 3157 }, { "epoch": 2.115204286671132, "grad_norm": 2.511603636064236, "learning_rate": 2.685437213463274e-06, "loss": 1.1235, "step": 3158 }, { "epoch": 2.115874079035499, "grad_norm": 2.5381294237962373, "learning_rate": 2.683979730079369e-06, "loss": 0.9832, "step": 3159 }, { "epoch": 2.116543871399866, "grad_norm": 2.3661769235756323, "learning_rate": 2.6825221838209865e-06, "loss": 1.1069, "step": 3160 }, { "epoch": 2.117213663764233, "grad_norm": 2.2466410960810714, "learning_rate": 2.6810645751862392e-06, "loss": 1.2091, "step": 3161 }, { "epoch": 2.1178834561286, "grad_norm": 3.821718518963136, "learning_rate": 2.679606904673259e-06, "loss": 0.9029, "step": 3162 }, { "epoch": 2.118553248492967, "grad_norm": 3.7810726036766242, "learning_rate": 2.6781491727802007e-06, "loss": 0.9904, "step": 3163 }, { "epoch": 2.1192230408573343, "grad_norm": 2.187025585681854, "learning_rate": 2.6766913800052395e-06, "loss": 1.0819, "step": 3164 }, { "epoch": 2.119892833221701, "grad_norm": 2.5597617874833523, "learning_rate": 2.6752335268465714e-06, "loss": 1.2195, "step": 3165 }, { "epoch": 2.1205626255860683, "grad_norm": 2.408395132699964, "learning_rate": 2.673775613802412e-06, "loss": 1.1033, "step": 3166 }, { "epoch": 2.121232417950435, "grad_norm": 2.7079577947015285, "learning_rate": 2.6723176413710006e-06, "loss": 0.8143, "step": 3167 }, { "epoch": 2.1219022103148024, "grad_norm": 2.3705160444849334, "learning_rate": 2.6708596100505924e-06, "loss": 1.274, "step": 3168 }, { "epoch": 2.1225720026791697, "grad_norm": 2.3167415325332468, "learning_rate": 2.6694015203394656e-06, "loss": 1.0022, "step": 3169 }, { "epoch": 2.1232417950435365, "grad_norm": 2.411119867209397, "learning_rate": 2.6679433727359176e-06, "loss": 1.2055, "step": 3170 }, { "epoch": 2.1239115874079038, "grad_norm": 2.30947562391941, "learning_rate": 2.6664851677382658e-06, "loss": 1.2541, "step": 3171 }, { "epoch": 2.1245813797722706, "grad_norm": 2.7706960855880554, "learning_rate": 2.6650269058448476e-06, "loss": 0.9785, "step": 3172 }, { "epoch": 2.125251172136638, "grad_norm": 2.4984595817122517, "learning_rate": 2.663568587554018e-06, "loss": 1.1723, "step": 3173 }, { "epoch": 2.1259209645010047, "grad_norm": 2.387080381577696, "learning_rate": 2.662110213364153e-06, "loss": 1.1643, "step": 3174 }, { "epoch": 2.126590756865372, "grad_norm": 2.283977973121114, "learning_rate": 2.660651783773648e-06, "loss": 1.1772, "step": 3175 }, { "epoch": 2.1272605492297387, "grad_norm": 2.6024200859118323, "learning_rate": 2.6591932992809153e-06, "loss": 1.0907, "step": 3176 }, { "epoch": 2.127930341594106, "grad_norm": 2.4059292536967405, "learning_rate": 2.6577347603843885e-06, "loss": 1.0954, "step": 3177 }, { "epoch": 2.128600133958473, "grad_norm": 2.3871275067014515, "learning_rate": 2.6562761675825185e-06, "loss": 1.1296, "step": 3178 }, { "epoch": 2.12926992632284, "grad_norm": 2.382819302284571, "learning_rate": 2.6548175213737727e-06, "loss": 1.2389, "step": 3179 }, { "epoch": 2.129939718687207, "grad_norm": 2.333474265184559, "learning_rate": 2.6533588222566415e-06, "loss": 1.1443, "step": 3180 }, { "epoch": 2.130609511051574, "grad_norm": 2.2028374261571204, "learning_rate": 2.6519000707296283e-06, "loss": 1.0712, "step": 3181 }, { "epoch": 2.131279303415941, "grad_norm": 2.900059711385245, "learning_rate": 2.6504412672912594e-06, "loss": 1.04, "step": 3182 }, { "epoch": 2.1319490957803082, "grad_norm": 2.5103222749435004, "learning_rate": 2.6489824124400727e-06, "loss": 1.1914, "step": 3183 }, { "epoch": 2.132618888144675, "grad_norm": 2.515258333217038, "learning_rate": 2.6475235066746304e-06, "loss": 1.0981, "step": 3184 }, { "epoch": 2.1332886805090423, "grad_norm": 2.111574128303949, "learning_rate": 2.6460645504935063e-06, "loss": 0.9538, "step": 3185 }, { "epoch": 2.133958472873409, "grad_norm": 2.023754072979498, "learning_rate": 2.6446055443952957e-06, "loss": 0.9487, "step": 3186 }, { "epoch": 2.1346282652377764, "grad_norm": 3.9373819127397, "learning_rate": 2.643146488878607e-06, "loss": 1.2653, "step": 3187 }, { "epoch": 2.135298057602143, "grad_norm": 2.5886687515898794, "learning_rate": 2.6416873844420703e-06, "loss": 1.0482, "step": 3188 }, { "epoch": 2.1359678499665105, "grad_norm": 2.5027944961173594, "learning_rate": 2.640228231584329e-06, "loss": 1.009, "step": 3189 }, { "epoch": 2.1366376423308773, "grad_norm": 2.5143895109793406, "learning_rate": 2.638769030804043e-06, "loss": 1.1917, "step": 3190 }, { "epoch": 2.1373074346952445, "grad_norm": 2.6783607104384712, "learning_rate": 2.6373097825998906e-06, "loss": 0.9945, "step": 3191 }, { "epoch": 2.1379772270596114, "grad_norm": 2.276437354304439, "learning_rate": 2.6358504874705638e-06, "loss": 1.2863, "step": 3192 }, { "epoch": 2.1386470194239786, "grad_norm": 3.7905921727802685, "learning_rate": 2.6343911459147732e-06, "loss": 0.9637, "step": 3193 }, { "epoch": 2.1393168117883454, "grad_norm": 2.5051301105446235, "learning_rate": 2.6329317584312436e-06, "loss": 1.2639, "step": 3194 }, { "epoch": 2.1399866041527127, "grad_norm": 3.9217158191426886, "learning_rate": 2.6314723255187158e-06, "loss": 1.1289, "step": 3195 }, { "epoch": 2.1406563965170795, "grad_norm": 3.773538264447804, "learning_rate": 2.6300128476759466e-06, "loss": 0.9031, "step": 3196 }, { "epoch": 2.1413261888814468, "grad_norm": 2.215971460344265, "learning_rate": 2.6285533254017076e-06, "loss": 1.0217, "step": 3197 }, { "epoch": 2.141995981245814, "grad_norm": 6.278525463356619, "learning_rate": 2.6270937591947855e-06, "loss": 1.0797, "step": 3198 }, { "epoch": 2.142665773610181, "grad_norm": 2.398324843413802, "learning_rate": 2.6256341495539833e-06, "loss": 1.1572, "step": 3199 }, { "epoch": 2.1433355659745477, "grad_norm": 2.9315576533146266, "learning_rate": 2.624174496978117e-06, "loss": 1.0362, "step": 3200 }, { "epoch": 2.144005358338915, "grad_norm": 2.465196611845113, "learning_rate": 2.6227148019660183e-06, "loss": 0.9314, "step": 3201 }, { "epoch": 2.144675150703282, "grad_norm": 2.715979277261311, "learning_rate": 2.621255065016533e-06, "loss": 1.098, "step": 3202 }, { "epoch": 2.145344943067649, "grad_norm": 2.6903114497321763, "learning_rate": 2.619795286628522e-06, "loss": 1.2735, "step": 3203 }, { "epoch": 2.1460147354320163, "grad_norm": 3.8021108440231757, "learning_rate": 2.6183354673008597e-06, "loss": 0.9381, "step": 3204 }, { "epoch": 2.146684527796383, "grad_norm": 2.391252571597186, "learning_rate": 2.6168756075324348e-06, "loss": 1.0573, "step": 3205 }, { "epoch": 2.1473543201607503, "grad_norm": 2.2084260565528466, "learning_rate": 2.615415707822149e-06, "loss": 1.0167, "step": 3206 }, { "epoch": 2.148024112525117, "grad_norm": 2.3244988016760315, "learning_rate": 2.6139557686689192e-06, "loss": 1.0763, "step": 3207 }, { "epoch": 2.1486939048894844, "grad_norm": 2.9060446841661816, "learning_rate": 2.6124957905716746e-06, "loss": 1.072, "step": 3208 }, { "epoch": 2.1493636972538512, "grad_norm": 2.695234618130845, "learning_rate": 2.6110357740293574e-06, "loss": 1.0455, "step": 3209 }, { "epoch": 2.1500334896182185, "grad_norm": 2.4709164684262315, "learning_rate": 2.609575719540924e-06, "loss": 0.9768, "step": 3210 }, { "epoch": 2.1507032819825853, "grad_norm": 2.4356918692655403, "learning_rate": 2.6081156276053432e-06, "loss": 1.3131, "step": 3211 }, { "epoch": 2.1513730743469526, "grad_norm": 4.596171229303035, "learning_rate": 2.6066554987215964e-06, "loss": 1.0779, "step": 3212 }, { "epoch": 2.1520428667113194, "grad_norm": 5.165045955849474, "learning_rate": 2.605195333388679e-06, "loss": 0.8279, "step": 3213 }, { "epoch": 2.1527126590756867, "grad_norm": 2.87689644841092, "learning_rate": 2.603735132105596e-06, "loss": 1.0813, "step": 3214 }, { "epoch": 2.1533824514400535, "grad_norm": 5.316648147407465, "learning_rate": 2.602274895371368e-06, "loss": 1.0463, "step": 3215 }, { "epoch": 2.1540522438044207, "grad_norm": 2.7350389271349402, "learning_rate": 2.600814623685025e-06, "loss": 1.0265, "step": 3216 }, { "epoch": 2.1547220361687875, "grad_norm": 2.2205149286972734, "learning_rate": 2.599354317545612e-06, "loss": 1.0217, "step": 3217 }, { "epoch": 2.155391828533155, "grad_norm": 2.173450019426964, "learning_rate": 2.5978939774521816e-06, "loss": 0.8872, "step": 3218 }, { "epoch": 2.1560616208975216, "grad_norm": 2.421456197194115, "learning_rate": 2.5964336039038022e-06, "loss": 1.1588, "step": 3219 }, { "epoch": 2.156731413261889, "grad_norm": 2.9394131776289445, "learning_rate": 2.5949731973995503e-06, "loss": 0.9016, "step": 3220 }, { "epoch": 2.1574012056262557, "grad_norm": 2.1947252362496843, "learning_rate": 2.5935127584385163e-06, "loss": 1.1199, "step": 3221 }, { "epoch": 2.158070997990623, "grad_norm": 3.0358226033732656, "learning_rate": 2.5920522875197982e-06, "loss": 1.2649, "step": 3222 }, { "epoch": 2.15874079035499, "grad_norm": 2.1972209224249166, "learning_rate": 2.590591785142511e-06, "loss": 1.2091, "step": 3223 }, { "epoch": 2.159410582719357, "grad_norm": 2.426563543892956, "learning_rate": 2.589131251805773e-06, "loss": 1.1637, "step": 3224 }, { "epoch": 2.160080375083724, "grad_norm": 2.695501063848771, "learning_rate": 2.587670688008719e-06, "loss": 0.9623, "step": 3225 }, { "epoch": 2.160750167448091, "grad_norm": 5.455904179888782, "learning_rate": 2.586210094250491e-06, "loss": 0.9788, "step": 3226 }, { "epoch": 2.161419959812458, "grad_norm": 2.3046820536295978, "learning_rate": 2.5847494710302433e-06, "loss": 1.0651, "step": 3227 }, { "epoch": 2.162089752176825, "grad_norm": 2.1976168841057198, "learning_rate": 2.5832888188471365e-06, "loss": 1.1979, "step": 3228 }, { "epoch": 2.162759544541192, "grad_norm": 2.282679836661058, "learning_rate": 2.5818281382003472e-06, "loss": 1.0736, "step": 3229 }, { "epoch": 2.1634293369055593, "grad_norm": 2.348249103816069, "learning_rate": 2.5803674295890553e-06, "loss": 1.0843, "step": 3230 }, { "epoch": 2.1640991292699265, "grad_norm": 2.1789229343092966, "learning_rate": 2.578906693512456e-06, "loss": 1.2215, "step": 3231 }, { "epoch": 2.1647689216342934, "grad_norm": 2.3861689663764283, "learning_rate": 2.577445930469748e-06, "loss": 1.0341, "step": 3232 }, { "epoch": 2.1654387139986606, "grad_norm": 2.271027146166203, "learning_rate": 2.5759851409601447e-06, "loss": 1.2851, "step": 3233 }, { "epoch": 2.1661085063630274, "grad_norm": 2.4035038511655635, "learning_rate": 2.5745243254828657e-06, "loss": 1.2593, "step": 3234 }, { "epoch": 2.1667782987273947, "grad_norm": 2.427462005321533, "learning_rate": 2.5730634845371392e-06, "loss": 1.2373, "step": 3235 }, { "epoch": 2.1674480910917615, "grad_norm": 2.7528406735029716, "learning_rate": 2.571602618622204e-06, "loss": 1.0981, "step": 3236 }, { "epoch": 2.1681178834561288, "grad_norm": 2.5538188810015923, "learning_rate": 2.570141728237306e-06, "loss": 1.2382, "step": 3237 }, { "epoch": 2.1687876758204956, "grad_norm": 2.9456896394209924, "learning_rate": 2.568680813881699e-06, "loss": 1.1383, "step": 3238 }, { "epoch": 2.169457468184863, "grad_norm": 2.429242594515706, "learning_rate": 2.567219876054646e-06, "loss": 1.0633, "step": 3239 }, { "epoch": 2.1701272605492297, "grad_norm": 2.5957516022067204, "learning_rate": 2.5657589152554183e-06, "loss": 1.0236, "step": 3240 }, { "epoch": 2.170797052913597, "grad_norm": 3.248190267396083, "learning_rate": 2.5642979319832943e-06, "loss": 0.9711, "step": 3241 }, { "epoch": 2.1714668452779637, "grad_norm": 2.746193975338019, "learning_rate": 2.56283692673756e-06, "loss": 1.3419, "step": 3242 }, { "epoch": 2.172136637642331, "grad_norm": 2.3597572153393207, "learning_rate": 2.56137590001751e-06, "loss": 1.1116, "step": 3243 }, { "epoch": 2.172806430006698, "grad_norm": 2.575383576205654, "learning_rate": 2.5599148523224438e-06, "loss": 1.1135, "step": 3244 }, { "epoch": 2.173476222371065, "grad_norm": 2.127314929905549, "learning_rate": 2.5584537841516716e-06, "loss": 0.8038, "step": 3245 }, { "epoch": 2.174146014735432, "grad_norm": 2.444435478395575, "learning_rate": 2.556992696004508e-06, "loss": 1.04, "step": 3246 }, { "epoch": 2.174815807099799, "grad_norm": 2.488271251326648, "learning_rate": 2.5555315883802743e-06, "loss": 1.0453, "step": 3247 }, { "epoch": 2.175485599464166, "grad_norm": 2.184008442164058, "learning_rate": 2.5540704617783006e-06, "loss": 1.0256, "step": 3248 }, { "epoch": 2.1761553918285332, "grad_norm": 3.286565446170577, "learning_rate": 2.5526093166979215e-06, "loss": 1.0519, "step": 3249 }, { "epoch": 2.1768251841929, "grad_norm": 2.4556582967133087, "learning_rate": 2.551148153638479e-06, "loss": 1.1559, "step": 3250 }, { "epoch": 2.1774949765572673, "grad_norm": 2.813788476541885, "learning_rate": 2.54968697309932e-06, "loss": 1.0122, "step": 3251 }, { "epoch": 2.178164768921634, "grad_norm": 3.0542214174638675, "learning_rate": 2.5482257755797996e-06, "loss": 1.0258, "step": 3252 }, { "epoch": 2.1788345612860014, "grad_norm": 2.7501973827984143, "learning_rate": 2.5467645615792757e-06, "loss": 0.75, "step": 3253 }, { "epoch": 2.179504353650368, "grad_norm": 3.0294731356562856, "learning_rate": 2.5453033315971154e-06, "loss": 0.9533, "step": 3254 }, { "epoch": 2.1801741460147355, "grad_norm": 2.4403466476104074, "learning_rate": 2.543842086132687e-06, "loss": 1.1886, "step": 3255 }, { "epoch": 2.1808439383791023, "grad_norm": 2.821617714389393, "learning_rate": 2.542380825685369e-06, "loss": 1.1751, "step": 3256 }, { "epoch": 2.1815137307434695, "grad_norm": 2.4612584408275553, "learning_rate": 2.540919550754539e-06, "loss": 1.087, "step": 3257 }, { "epoch": 2.1821835231078364, "grad_norm": 2.897006669876616, "learning_rate": 2.5394582618395868e-06, "loss": 1.1984, "step": 3258 }, { "epoch": 2.1828533154722036, "grad_norm": 2.5916822400779087, "learning_rate": 2.537996959439901e-06, "loss": 1.1998, "step": 3259 }, { "epoch": 2.183523107836571, "grad_norm": 2.759622611585917, "learning_rate": 2.536535644054877e-06, "loss": 0.9098, "step": 3260 }, { "epoch": 2.1841929002009377, "grad_norm": 2.3847157134044554, "learning_rate": 2.5350743161839147e-06, "loss": 1.2007, "step": 3261 }, { "epoch": 2.1848626925653045, "grad_norm": 2.5537121297830323, "learning_rate": 2.5336129763264196e-06, "loss": 0.9838, "step": 3262 }, { "epoch": 2.185532484929672, "grad_norm": 2.693958865999435, "learning_rate": 2.5321516249817967e-06, "loss": 1.1377, "step": 3263 }, { "epoch": 2.186202277294039, "grad_norm": 3.5196449554241305, "learning_rate": 2.5306902626494623e-06, "loss": 1.307, "step": 3264 }, { "epoch": 2.186872069658406, "grad_norm": 2.4614629071050476, "learning_rate": 2.529228889828828e-06, "loss": 1.1089, "step": 3265 }, { "epoch": 2.187541862022773, "grad_norm": 4.0809077921853465, "learning_rate": 2.5277675070193167e-06, "loss": 1.1394, "step": 3266 }, { "epoch": 2.18821165438714, "grad_norm": 2.3460851197151205, "learning_rate": 2.526306114720349e-06, "loss": 1.3152, "step": 3267 }, { "epoch": 2.188881446751507, "grad_norm": 2.562553942494308, "learning_rate": 2.5248447134313526e-06, "loss": 0.9797, "step": 3268 }, { "epoch": 2.189551239115874, "grad_norm": 2.1910928886269105, "learning_rate": 2.523383303651754e-06, "loss": 1.1374, "step": 3269 }, { "epoch": 2.1902210314802413, "grad_norm": 2.6626946118786026, "learning_rate": 2.5219218858809893e-06, "loss": 1.017, "step": 3270 }, { "epoch": 2.190890823844608, "grad_norm": 3.605611454426158, "learning_rate": 2.520460460618489e-06, "loss": 0.9923, "step": 3271 }, { "epoch": 2.1915606162089754, "grad_norm": 2.4440361556629657, "learning_rate": 2.5189990283636946e-06, "loss": 1.1762, "step": 3272 }, { "epoch": 2.192230408573342, "grad_norm": 2.4974538544165967, "learning_rate": 2.517537589616042e-06, "loss": 1.257, "step": 3273 }, { "epoch": 2.1929002009377094, "grad_norm": 2.307264123418192, "learning_rate": 2.5160761448749764e-06, "loss": 1.0603, "step": 3274 }, { "epoch": 2.1935699933020762, "grad_norm": 2.19413592162699, "learning_rate": 2.514614694639939e-06, "loss": 0.9562, "step": 3275 }, { "epoch": 2.1942397856664435, "grad_norm": 3.7440437359873613, "learning_rate": 2.513153239410378e-06, "loss": 1.0582, "step": 3276 }, { "epoch": 2.1949095780308103, "grad_norm": 2.112965639781095, "learning_rate": 2.511691779685739e-06, "loss": 1.0471, "step": 3277 }, { "epoch": 2.1955793703951776, "grad_norm": 2.3904872638003933, "learning_rate": 2.510230315965473e-06, "loss": 1.1994, "step": 3278 }, { "epoch": 2.1962491627595444, "grad_norm": 2.239820505344623, "learning_rate": 2.5087688487490296e-06, "loss": 1.1433, "step": 3279 }, { "epoch": 2.1969189551239117, "grad_norm": 2.2887902652583842, "learning_rate": 2.5073073785358606e-06, "loss": 1.2024, "step": 3280 }, { "epoch": 2.1975887474882785, "grad_norm": 2.8411621350725795, "learning_rate": 2.505845905825418e-06, "loss": 1.1179, "step": 3281 }, { "epoch": 2.1982585398526457, "grad_norm": 2.5616338764930733, "learning_rate": 2.5043844311171566e-06, "loss": 1.1307, "step": 3282 }, { "epoch": 2.1989283322170126, "grad_norm": 2.366876115466192, "learning_rate": 2.5029229549105298e-06, "loss": 1.1022, "step": 3283 }, { "epoch": 2.19959812458138, "grad_norm": 2.4426650435426853, "learning_rate": 2.5014614777049927e-06, "loss": 1.1853, "step": 3284 }, { "epoch": 2.2002679169457466, "grad_norm": 2.992577470940068, "learning_rate": 2.5e-06, "loss": 1.0947, "step": 3285 }, { "epoch": 2.200937709310114, "grad_norm": 2.3702157964623765, "learning_rate": 2.4985385222950077e-06, "loss": 1.1082, "step": 3286 }, { "epoch": 2.2016075016744807, "grad_norm": 2.4159928483237803, "learning_rate": 2.4970770450894707e-06, "loss": 1.2505, "step": 3287 }, { "epoch": 2.202277294038848, "grad_norm": 2.279692646930696, "learning_rate": 2.4956155688828442e-06, "loss": 1.1944, "step": 3288 }, { "epoch": 2.2029470864032152, "grad_norm": 2.437467709821535, "learning_rate": 2.4941540941745827e-06, "loss": 1.0671, "step": 3289 }, { "epoch": 2.203616878767582, "grad_norm": 2.5293845816179683, "learning_rate": 2.4926926214641407e-06, "loss": 1.1988, "step": 3290 }, { "epoch": 2.204286671131949, "grad_norm": 2.5976970262308012, "learning_rate": 2.4912311512509708e-06, "loss": 1.0518, "step": 3291 }, { "epoch": 2.204956463496316, "grad_norm": 2.3373824862793025, "learning_rate": 2.4897696840345276e-06, "loss": 1.0754, "step": 3292 }, { "epoch": 2.2056262558606834, "grad_norm": 2.4591627042083184, "learning_rate": 2.4883082203142612e-06, "loss": 1.0531, "step": 3293 }, { "epoch": 2.20629604822505, "grad_norm": 2.4951727762070144, "learning_rate": 2.4868467605896225e-06, "loss": 1.2312, "step": 3294 }, { "epoch": 2.2069658405894175, "grad_norm": 2.5306137632118633, "learning_rate": 2.4853853053600613e-06, "loss": 1.1686, "step": 3295 }, { "epoch": 2.2076356329537843, "grad_norm": 2.199741912157699, "learning_rate": 2.483923855125025e-06, "loss": 1.1294, "step": 3296 }, { "epoch": 2.2083054253181515, "grad_norm": 2.6442443236705695, "learning_rate": 2.4824624103839585e-06, "loss": 0.955, "step": 3297 }, { "epoch": 2.2089752176825184, "grad_norm": 2.7952818878151047, "learning_rate": 2.4810009716363063e-06, "loss": 1.0713, "step": 3298 }, { "epoch": 2.2096450100468856, "grad_norm": 2.4715624835399095, "learning_rate": 2.479539539381511e-06, "loss": 1.0069, "step": 3299 }, { "epoch": 2.2103148024112524, "grad_norm": 2.3741449855301737, "learning_rate": 2.478078114119012e-06, "loss": 1.182, "step": 3300 }, { "epoch": 2.2109845947756197, "grad_norm": 5.191734970661536, "learning_rate": 2.4766166963482463e-06, "loss": 0.8756, "step": 3301 }, { "epoch": 2.2116543871399865, "grad_norm": 2.6252076675904013, "learning_rate": 2.4751552865686486e-06, "loss": 0.9215, "step": 3302 }, { "epoch": 2.212324179504354, "grad_norm": 2.7767780827425175, "learning_rate": 2.473693885279651e-06, "loss": 1.1987, "step": 3303 }, { "epoch": 2.2129939718687206, "grad_norm": 3.6579643787260054, "learning_rate": 2.4722324929806845e-06, "loss": 1.1543, "step": 3304 }, { "epoch": 2.213663764233088, "grad_norm": 2.3450367069001756, "learning_rate": 2.470771110171173e-06, "loss": 1.1525, "step": 3305 }, { "epoch": 2.2143335565974547, "grad_norm": 4.27576362938234, "learning_rate": 2.4693097373505385e-06, "loss": 1.0123, "step": 3306 }, { "epoch": 2.215003348961822, "grad_norm": 2.296026004871393, "learning_rate": 2.467848375018203e-06, "loss": 1.2202, "step": 3307 }, { "epoch": 2.2156731413261888, "grad_norm": 2.328640483677928, "learning_rate": 2.466387023673582e-06, "loss": 1.2072, "step": 3308 }, { "epoch": 2.216342933690556, "grad_norm": 4.41405802743263, "learning_rate": 2.4649256838160857e-06, "loss": 1.0588, "step": 3309 }, { "epoch": 2.217012726054923, "grad_norm": 2.923560264560979, "learning_rate": 2.4634643559451234e-06, "loss": 1.2325, "step": 3310 }, { "epoch": 2.21768251841929, "grad_norm": 2.810151199277657, "learning_rate": 2.4620030405600996e-06, "loss": 1.1887, "step": 3311 }, { "epoch": 2.218352310783657, "grad_norm": 2.3193263959856387, "learning_rate": 2.460541738160414e-06, "loss": 0.9452, "step": 3312 }, { "epoch": 2.219022103148024, "grad_norm": 2.38900222628456, "learning_rate": 2.4590804492454613e-06, "loss": 1.1919, "step": 3313 }, { "epoch": 2.219691895512391, "grad_norm": 2.3253010608526643, "learning_rate": 2.457619174314632e-06, "loss": 1.2509, "step": 3314 }, { "epoch": 2.2203616878767582, "grad_norm": 2.8341175983884406, "learning_rate": 2.4561579138673134e-06, "loss": 1.0232, "step": 3315 }, { "epoch": 2.221031480241125, "grad_norm": 2.3750481274666906, "learning_rate": 2.4546966684028863e-06, "loss": 1.0376, "step": 3316 }, { "epoch": 2.2217012726054923, "grad_norm": 2.278074179404438, "learning_rate": 2.4532354384207247e-06, "loss": 1.0752, "step": 3317 }, { "epoch": 2.222371064969859, "grad_norm": 2.3807300642982656, "learning_rate": 2.4517742244202013e-06, "loss": 1.1968, "step": 3318 }, { "epoch": 2.2230408573342264, "grad_norm": 2.2307595036308587, "learning_rate": 2.4503130269006803e-06, "loss": 1.1607, "step": 3319 }, { "epoch": 2.2237106496985932, "grad_norm": 2.119690088484262, "learning_rate": 2.4488518463615225e-06, "loss": 1.1942, "step": 3320 }, { "epoch": 2.2243804420629605, "grad_norm": 2.3535713500084463, "learning_rate": 2.4473906833020797e-06, "loss": 1.207, "step": 3321 }, { "epoch": 2.2250502344273277, "grad_norm": 2.4881335630928865, "learning_rate": 2.4459295382217e-06, "loss": 0.9336, "step": 3322 }, { "epoch": 2.2257200267916946, "grad_norm": 2.435081118269479, "learning_rate": 2.444468411619726e-06, "loss": 0.9617, "step": 3323 }, { "epoch": 2.226389819156062, "grad_norm": 2.854983358963258, "learning_rate": 2.4430073039954925e-06, "loss": 1.0922, "step": 3324 }, { "epoch": 2.2270596115204286, "grad_norm": 3.870099226163477, "learning_rate": 2.441546215848329e-06, "loss": 1.1032, "step": 3325 }, { "epoch": 2.227729403884796, "grad_norm": 2.5900321722602486, "learning_rate": 2.4400851476775566e-06, "loss": 1.1121, "step": 3326 }, { "epoch": 2.2283991962491627, "grad_norm": 2.2669171315025545, "learning_rate": 2.438624099982491e-06, "loss": 1.0833, "step": 3327 }, { "epoch": 2.22906898861353, "grad_norm": 2.882425290986003, "learning_rate": 2.4371630732624404e-06, "loss": 1.1925, "step": 3328 }, { "epoch": 2.229738780977897, "grad_norm": 2.958076987213801, "learning_rate": 2.4357020680167066e-06, "loss": 1.2202, "step": 3329 }, { "epoch": 2.230408573342264, "grad_norm": 2.3506895553120195, "learning_rate": 2.4342410847445825e-06, "loss": 1.2197, "step": 3330 }, { "epoch": 2.231078365706631, "grad_norm": 3.385381235450904, "learning_rate": 2.4327801239453547e-06, "loss": 1.0849, "step": 3331 }, { "epoch": 2.231748158070998, "grad_norm": 2.56063519674606, "learning_rate": 2.4313191861183015e-06, "loss": 1.0982, "step": 3332 }, { "epoch": 2.232417950435365, "grad_norm": 3.402368967839854, "learning_rate": 2.429858271762695e-06, "loss": 1.0461, "step": 3333 }, { "epoch": 2.233087742799732, "grad_norm": 2.2288171750505716, "learning_rate": 2.4283973813777962e-06, "loss": 1.1059, "step": 3334 }, { "epoch": 2.233757535164099, "grad_norm": 3.551609770871597, "learning_rate": 2.426936515462861e-06, "loss": 0.9776, "step": 3335 }, { "epoch": 2.2344273275284663, "grad_norm": 2.7377978967366032, "learning_rate": 2.4254756745171348e-06, "loss": 1.2222, "step": 3336 }, { "epoch": 2.235097119892833, "grad_norm": 3.4865149342684525, "learning_rate": 2.424014859039856e-06, "loss": 1.1117, "step": 3337 }, { "epoch": 2.2357669122572004, "grad_norm": 2.5786464394449413, "learning_rate": 2.422554069530253e-06, "loss": 1.2061, "step": 3338 }, { "epoch": 2.236436704621567, "grad_norm": 2.1884546124006596, "learning_rate": 2.421093306487545e-06, "loss": 1.0672, "step": 3339 }, { "epoch": 2.2371064969859344, "grad_norm": 2.2360855908295187, "learning_rate": 2.4196325704109447e-06, "loss": 1.1637, "step": 3340 }, { "epoch": 2.2377762893503013, "grad_norm": 2.704434620209425, "learning_rate": 2.418171861799654e-06, "loss": 1.1001, "step": 3341 }, { "epoch": 2.2384460817146685, "grad_norm": 2.6992794688761914, "learning_rate": 2.416711181152864e-06, "loss": 0.9337, "step": 3342 }, { "epoch": 2.2391158740790353, "grad_norm": 2.282646248039567, "learning_rate": 2.4152505289697575e-06, "loss": 1.0471, "step": 3343 }, { "epoch": 2.2397856664434026, "grad_norm": 3.2527563400902206, "learning_rate": 2.413789905749509e-06, "loss": 1.1823, "step": 3344 }, { "epoch": 2.2404554588077694, "grad_norm": 2.563147938690302, "learning_rate": 2.4123293119912817e-06, "loss": 1.1969, "step": 3345 }, { "epoch": 2.2411252511721367, "grad_norm": 2.580636234535093, "learning_rate": 2.4108687481942274e-06, "loss": 1.1659, "step": 3346 }, { "epoch": 2.2417950435365035, "grad_norm": 3.0623788868227226, "learning_rate": 2.4094082148574897e-06, "loss": 1.0621, "step": 3347 }, { "epoch": 2.2424648359008708, "grad_norm": 2.2314485970072657, "learning_rate": 2.4079477124802013e-06, "loss": 0.8356, "step": 3348 }, { "epoch": 2.2431346282652376, "grad_norm": 3.870181776268372, "learning_rate": 2.406487241561485e-06, "loss": 1.1161, "step": 3349 }, { "epoch": 2.243804420629605, "grad_norm": 2.5483961766964813, "learning_rate": 2.4050268026004505e-06, "loss": 1.2441, "step": 3350 }, { "epoch": 2.244474212993972, "grad_norm": 2.8039950968663034, "learning_rate": 2.403566396096198e-06, "loss": 1.3314, "step": 3351 }, { "epoch": 2.245144005358339, "grad_norm": 2.8444304477009394, "learning_rate": 2.4021060225478184e-06, "loss": 1.1519, "step": 3352 }, { "epoch": 2.2458137977227057, "grad_norm": 2.3116330645366134, "learning_rate": 2.400645682454389e-06, "loss": 1.0757, "step": 3353 }, { "epoch": 2.246483590087073, "grad_norm": 2.3080211097481023, "learning_rate": 2.3991853763149757e-06, "loss": 1.2055, "step": 3354 }, { "epoch": 2.2471533824514403, "grad_norm": 2.5720133555748377, "learning_rate": 2.3977251046286326e-06, "loss": 0.9266, "step": 3355 }, { "epoch": 2.247823174815807, "grad_norm": 2.1377924234146506, "learning_rate": 2.3962648678944045e-06, "loss": 1.0608, "step": 3356 }, { "epoch": 2.2484929671801743, "grad_norm": 2.286074833928326, "learning_rate": 2.3948046666113216e-06, "loss": 0.9978, "step": 3357 }, { "epoch": 2.249162759544541, "grad_norm": 2.569085627461364, "learning_rate": 2.3933445012784045e-06, "loss": 1.0974, "step": 3358 }, { "epoch": 2.2498325519089084, "grad_norm": 3.1153000519369582, "learning_rate": 2.3918843723946576e-06, "loss": 1.1123, "step": 3359 }, { "epoch": 2.2505023442732752, "grad_norm": 5.537701487938681, "learning_rate": 2.390424280459077e-06, "loss": 1.0201, "step": 3360 }, { "epoch": 2.2511721366376425, "grad_norm": 2.948916846487736, "learning_rate": 2.3889642259706426e-06, "loss": 1.1185, "step": 3361 }, { "epoch": 2.2518419290020093, "grad_norm": 2.689903465495599, "learning_rate": 2.3875042094283267e-06, "loss": 1.2325, "step": 3362 }, { "epoch": 2.2525117213663766, "grad_norm": 2.473670038052647, "learning_rate": 2.386044231331081e-06, "loss": 1.0389, "step": 3363 }, { "epoch": 2.2531815137307434, "grad_norm": 4.391270803959236, "learning_rate": 2.3845842921778513e-06, "loss": 1.1579, "step": 3364 }, { "epoch": 2.2538513060951106, "grad_norm": 2.917719698606036, "learning_rate": 2.3831243924675652e-06, "loss": 1.1287, "step": 3365 }, { "epoch": 2.2545210984594775, "grad_norm": 2.346388207381168, "learning_rate": 2.3816645326991415e-06, "loss": 1.1728, "step": 3366 }, { "epoch": 2.2551908908238447, "grad_norm": 2.2086729168743777, "learning_rate": 2.3802047133714783e-06, "loss": 1.0365, "step": 3367 }, { "epoch": 2.2558606831882115, "grad_norm": 2.258063586198074, "learning_rate": 2.3787449349834674e-06, "loss": 1.1915, "step": 3368 }, { "epoch": 2.256530475552579, "grad_norm": 2.073852722245648, "learning_rate": 2.377285198033982e-06, "loss": 1.0271, "step": 3369 }, { "epoch": 2.2572002679169456, "grad_norm": 2.371777647269687, "learning_rate": 2.375825503021884e-06, "loss": 1.2125, "step": 3370 }, { "epoch": 2.257870060281313, "grad_norm": 2.4928014302657595, "learning_rate": 2.3743658504460176e-06, "loss": 1.1197, "step": 3371 }, { "epoch": 2.2585398526456797, "grad_norm": 2.1740398443208155, "learning_rate": 2.372906240805215e-06, "loss": 0.9826, "step": 3372 }, { "epoch": 2.259209645010047, "grad_norm": 2.574111039306844, "learning_rate": 2.371446674598293e-06, "loss": 1.0803, "step": 3373 }, { "epoch": 2.2598794373744138, "grad_norm": 3.0945078233106864, "learning_rate": 2.3699871523240542e-06, "loss": 1.179, "step": 3374 }, { "epoch": 2.260549229738781, "grad_norm": 2.668568591410303, "learning_rate": 2.368527674481285e-06, "loss": 0.9765, "step": 3375 }, { "epoch": 2.261219022103148, "grad_norm": 2.4997551143668084, "learning_rate": 2.3670682415687573e-06, "loss": 1.0791, "step": 3376 }, { "epoch": 2.261888814467515, "grad_norm": 2.35478280400911, "learning_rate": 2.3656088540852268e-06, "loss": 1.304, "step": 3377 }, { "epoch": 2.262558606831882, "grad_norm": 2.2322174310933316, "learning_rate": 2.3641495125294367e-06, "loss": 1.1687, "step": 3378 }, { "epoch": 2.263228399196249, "grad_norm": 3.3846260319769343, "learning_rate": 2.3626902174001102e-06, "loss": 0.9986, "step": 3379 }, { "epoch": 2.2638981915606164, "grad_norm": 2.4232219916543536, "learning_rate": 2.3612309691959576e-06, "loss": 1.224, "step": 3380 }, { "epoch": 2.2645679839249833, "grad_norm": 2.302511179931817, "learning_rate": 2.359771768415671e-06, "loss": 1.1065, "step": 3381 }, { "epoch": 2.26523777628935, "grad_norm": 2.7509912985725826, "learning_rate": 2.35831261555793e-06, "loss": 1.072, "step": 3382 }, { "epoch": 2.2659075686537173, "grad_norm": 3.788645583313155, "learning_rate": 2.3568535111213932e-06, "loss": 1.0411, "step": 3383 }, { "epoch": 2.2665773610180846, "grad_norm": 2.2711542955725084, "learning_rate": 2.3553944556047055e-06, "loss": 1.257, "step": 3384 }, { "epoch": 2.2672471533824514, "grad_norm": 2.4694624921596646, "learning_rate": 2.353935449506494e-06, "loss": 1.0516, "step": 3385 }, { "epoch": 2.2679169457468187, "grad_norm": 2.572823003892697, "learning_rate": 2.352476493325371e-06, "loss": 1.2309, "step": 3386 }, { "epoch": 2.2685867381111855, "grad_norm": 2.2926776147648984, "learning_rate": 2.351017587559928e-06, "loss": 1.0117, "step": 3387 }, { "epoch": 2.2692565304755528, "grad_norm": 3.0909514252042323, "learning_rate": 2.3495587327087414e-06, "loss": 1.2909, "step": 3388 }, { "epoch": 2.2699263228399196, "grad_norm": 4.957691734201597, "learning_rate": 2.3480999292703717e-06, "loss": 1.0765, "step": 3389 }, { "epoch": 2.270596115204287, "grad_norm": 3.5245372891941926, "learning_rate": 2.3466411777433594e-06, "loss": 0.9199, "step": 3390 }, { "epoch": 2.2712659075686537, "grad_norm": 2.4825689238162028, "learning_rate": 2.345182478626228e-06, "loss": 1.0123, "step": 3391 }, { "epoch": 2.271935699933021, "grad_norm": 2.5637685675280433, "learning_rate": 2.343723832417483e-06, "loss": 0.9911, "step": 3392 }, { "epoch": 2.2726054922973877, "grad_norm": 2.55266846752766, "learning_rate": 2.342265239615612e-06, "loss": 1.3193, "step": 3393 }, { "epoch": 2.273275284661755, "grad_norm": 2.556193451582239, "learning_rate": 2.340806700719085e-06, "loss": 1.0595, "step": 3394 }, { "epoch": 2.273945077026122, "grad_norm": 2.521556218008919, "learning_rate": 2.3393482162263537e-06, "loss": 1.1665, "step": 3395 }, { "epoch": 2.274614869390489, "grad_norm": 3.815264274056913, "learning_rate": 2.337889786635848e-06, "loss": 1.0439, "step": 3396 }, { "epoch": 2.275284661754856, "grad_norm": 2.3492118657184875, "learning_rate": 2.336431412445983e-06, "loss": 1.2015, "step": 3397 }, { "epoch": 2.275954454119223, "grad_norm": 3.1751133282710584, "learning_rate": 2.334973094155153e-06, "loss": 1.1207, "step": 3398 }, { "epoch": 2.27662424648359, "grad_norm": 2.7733786877698594, "learning_rate": 2.333514832261735e-06, "loss": 1.1563, "step": 3399 }, { "epoch": 2.2772940388479572, "grad_norm": 2.4169255774002134, "learning_rate": 2.332056627264083e-06, "loss": 1.1203, "step": 3400 }, { "epoch": 2.277963831212324, "grad_norm": 2.231021498392203, "learning_rate": 2.330598479660535e-06, "loss": 1.074, "step": 3401 }, { "epoch": 2.2786336235766913, "grad_norm": 2.3780817804768417, "learning_rate": 2.329140389949408e-06, "loss": 1.1524, "step": 3402 }, { "epoch": 2.279303415941058, "grad_norm": 3.349029488538457, "learning_rate": 2.327682358629001e-06, "loss": 1.0872, "step": 3403 }, { "epoch": 2.2799732083054254, "grad_norm": 3.7079999111646553, "learning_rate": 2.3262243861975882e-06, "loss": 0.9292, "step": 3404 }, { "epoch": 2.280643000669792, "grad_norm": 2.358076841576556, "learning_rate": 2.3247664731534294e-06, "loss": 1.2266, "step": 3405 }, { "epoch": 2.2813127930341595, "grad_norm": 2.387636902756746, "learning_rate": 2.323308619994761e-06, "loss": 1.0624, "step": 3406 }, { "epoch": 2.2819825853985263, "grad_norm": 5.601356328420814, "learning_rate": 2.3218508272198006e-06, "loss": 1.2007, "step": 3407 }, { "epoch": 2.2826523777628935, "grad_norm": 2.7239315392029626, "learning_rate": 2.320393095326742e-06, "loss": 0.9672, "step": 3408 }, { "epoch": 2.283322170127261, "grad_norm": 2.9123788878069528, "learning_rate": 2.318935424813762e-06, "loss": 1.2229, "step": 3409 }, { "epoch": 2.2839919624916276, "grad_norm": 2.2438782993239283, "learning_rate": 2.317477816179014e-06, "loss": 1.0948, "step": 3410 }, { "epoch": 2.2846617548559944, "grad_norm": 2.4408282891001285, "learning_rate": 2.3160202699206326e-06, "loss": 1.2147, "step": 3411 }, { "epoch": 2.2853315472203617, "grad_norm": 3.271123251761047, "learning_rate": 2.314562786536727e-06, "loss": 1.1467, "step": 3412 }, { "epoch": 2.286001339584729, "grad_norm": 3.9215765254520814, "learning_rate": 2.313105366525389e-06, "loss": 0.8838, "step": 3413 }, { "epoch": 2.2866711319490958, "grad_norm": 2.414405082190119, "learning_rate": 2.3116480103846863e-06, "loss": 1.2618, "step": 3414 }, { "epoch": 2.2873409243134626, "grad_norm": 2.25672328338871, "learning_rate": 2.3101907186126673e-06, "loss": 0.916, "step": 3415 }, { "epoch": 2.28801071667783, "grad_norm": 2.3320839882547926, "learning_rate": 2.308733491707355e-06, "loss": 1.1327, "step": 3416 }, { "epoch": 2.288680509042197, "grad_norm": 3.755324413943816, "learning_rate": 2.3072763301667535e-06, "loss": 0.9869, "step": 3417 }, { "epoch": 2.289350301406564, "grad_norm": 2.7825439521743216, "learning_rate": 2.305819234488841e-06, "loss": 1.0524, "step": 3418 }, { "epoch": 2.290020093770931, "grad_norm": 2.6822474202104423, "learning_rate": 2.304362205171578e-06, "loss": 1.244, "step": 3419 }, { "epoch": 2.290689886135298, "grad_norm": 2.7995977854175473, "learning_rate": 2.3029052427128966e-06, "loss": 1.2152, "step": 3420 }, { "epoch": 2.2913596784996653, "grad_norm": 2.6013945262641207, "learning_rate": 2.301448347610711e-06, "loss": 1.041, "step": 3421 }, { "epoch": 2.292029470864032, "grad_norm": 4.597475592137748, "learning_rate": 2.299991520362909e-06, "loss": 0.9069, "step": 3422 }, { "epoch": 2.2926992632283993, "grad_norm": 7.488020202619901, "learning_rate": 2.2985347614673588e-06, "loss": 0.977, "step": 3423 }, { "epoch": 2.293369055592766, "grad_norm": 3.402085270878958, "learning_rate": 2.2970780714219005e-06, "loss": 0.9471, "step": 3424 }, { "epoch": 2.2940388479571334, "grad_norm": 2.443543140841833, "learning_rate": 2.295621450724355e-06, "loss": 1.0895, "step": 3425 }, { "epoch": 2.2947086403215002, "grad_norm": 3.677888517094213, "learning_rate": 2.2941648998725153e-06, "loss": 1.0658, "step": 3426 }, { "epoch": 2.2953784326858675, "grad_norm": 2.3742011183109426, "learning_rate": 2.2927084193641566e-06, "loss": 1.2333, "step": 3427 }, { "epoch": 2.2960482250502343, "grad_norm": 2.4839771634549077, "learning_rate": 2.291252009697024e-06, "loss": 1.2537, "step": 3428 }, { "epoch": 2.2967180174146016, "grad_norm": 2.3833753977844188, "learning_rate": 2.2897956713688402e-06, "loss": 1.0467, "step": 3429 }, { "epoch": 2.2973878097789684, "grad_norm": 2.5402697318187077, "learning_rate": 2.2883394048773063e-06, "loss": 1.0136, "step": 3430 }, { "epoch": 2.2980576021433357, "grad_norm": 2.3040207795136256, "learning_rate": 2.286883210720095e-06, "loss": 1.0248, "step": 3431 }, { "epoch": 2.2987273945077025, "grad_norm": 2.2644119016463207, "learning_rate": 2.2854270893948568e-06, "loss": 0.9704, "step": 3432 }, { "epoch": 2.2993971868720697, "grad_norm": 3.4455570537968736, "learning_rate": 2.2839710413992155e-06, "loss": 1.0025, "step": 3433 }, { "epoch": 2.3000669792364365, "grad_norm": 3.0448174205068375, "learning_rate": 2.2825150672307715e-06, "loss": 1.2573, "step": 3434 }, { "epoch": 2.300736771600804, "grad_norm": 2.3604131454489035, "learning_rate": 2.2810591673871e-06, "loss": 0.9186, "step": 3435 }, { "epoch": 2.3014065639651706, "grad_norm": 2.5906617537190204, "learning_rate": 2.2796033423657486e-06, "loss": 1.1111, "step": 3436 }, { "epoch": 2.302076356329538, "grad_norm": 3.27281779468534, "learning_rate": 2.278147592664241e-06, "loss": 1.0375, "step": 3437 }, { "epoch": 2.3027461486939047, "grad_norm": 3.052926541012571, "learning_rate": 2.2766919187800753e-06, "loss": 1.1505, "step": 3438 }, { "epoch": 2.303415941058272, "grad_norm": 2.8946883362706752, "learning_rate": 2.2752363212107238e-06, "loss": 1.0097, "step": 3439 }, { "epoch": 2.304085733422639, "grad_norm": 2.4630438984591176, "learning_rate": 2.2737808004536323e-06, "loss": 1.3203, "step": 3440 }, { "epoch": 2.304755525787006, "grad_norm": 2.3878189933029117, "learning_rate": 2.2723253570062187e-06, "loss": 1.0594, "step": 3441 }, { "epoch": 2.3054253181513733, "grad_norm": 2.216674603720947, "learning_rate": 2.270869991365878e-06, "loss": 1.1561, "step": 3442 }, { "epoch": 2.30609511051574, "grad_norm": 2.5863971508608805, "learning_rate": 2.269414704029975e-06, "loss": 1.222, "step": 3443 }, { "epoch": 2.306764902880107, "grad_norm": 2.3639912799834573, "learning_rate": 2.267959495495852e-06, "loss": 1.2109, "step": 3444 }, { "epoch": 2.307434695244474, "grad_norm": 2.6717964650942863, "learning_rate": 2.266504366260819e-06, "loss": 1.0398, "step": 3445 }, { "epoch": 2.3081044876088415, "grad_norm": 2.1250934363718335, "learning_rate": 2.265049316822163e-06, "loss": 0.9394, "step": 3446 }, { "epoch": 2.3087742799732083, "grad_norm": 2.421309354471688, "learning_rate": 2.2635943476771416e-06, "loss": 1.1825, "step": 3447 }, { "epoch": 2.3094440723375755, "grad_norm": 2.3537809112874353, "learning_rate": 2.2621394593229885e-06, "loss": 1.1736, "step": 3448 }, { "epoch": 2.3101138647019424, "grad_norm": 3.034706004593537, "learning_rate": 2.260684652256903e-06, "loss": 0.9949, "step": 3449 }, { "epoch": 2.3107836570663096, "grad_norm": 2.9918383568759914, "learning_rate": 2.2592299269760634e-06, "loss": 1.1163, "step": 3450 }, { "epoch": 2.3114534494306764, "grad_norm": 2.3285583376784382, "learning_rate": 2.2577752839776167e-06, "loss": 0.9929, "step": 3451 }, { "epoch": 2.3121232417950437, "grad_norm": 2.529723660046189, "learning_rate": 2.2563207237586834e-06, "loss": 1.1621, "step": 3452 }, { "epoch": 2.3127930341594105, "grad_norm": 3.177786403220169, "learning_rate": 2.2548662468163525e-06, "loss": 1.0329, "step": 3453 }, { "epoch": 2.3134628265237778, "grad_norm": 2.4609053945756787, "learning_rate": 2.253411853647689e-06, "loss": 1.0881, "step": 3454 }, { "epoch": 2.3141326188881446, "grad_norm": 2.341850333738202, "learning_rate": 2.251957544749725e-06, "loss": 1.1946, "step": 3455 }, { "epoch": 2.314802411252512, "grad_norm": 2.4034502920192518, "learning_rate": 2.2505033206194683e-06, "loss": 0.9655, "step": 3456 }, { "epoch": 2.3154722036168787, "grad_norm": 2.6516292273493436, "learning_rate": 2.2490491817538936e-06, "loss": 1.129, "step": 3457 }, { "epoch": 2.316141995981246, "grad_norm": 2.8098508651686913, "learning_rate": 2.2475951286499483e-06, "loss": 1.0671, "step": 3458 }, { "epoch": 2.3168117883456127, "grad_norm": 2.6199157600314456, "learning_rate": 2.24614116180455e-06, "loss": 1.2311, "step": 3459 }, { "epoch": 2.31748158070998, "grad_norm": 2.921890069002517, "learning_rate": 2.244687281714589e-06, "loss": 1.113, "step": 3460 }, { "epoch": 2.318151373074347, "grad_norm": 2.4220211164613232, "learning_rate": 2.2432334888769225e-06, "loss": 1.0258, "step": 3461 }, { "epoch": 2.318821165438714, "grad_norm": 2.5827885960262735, "learning_rate": 2.2417797837883796e-06, "loss": 1.1378, "step": 3462 }, { "epoch": 2.319490957803081, "grad_norm": 2.518623948950109, "learning_rate": 2.240326166945759e-06, "loss": 1.2012, "step": 3463 }, { "epoch": 2.320160750167448, "grad_norm": 3.473404278794593, "learning_rate": 2.2388726388458313e-06, "loss": 1.0196, "step": 3464 }, { "epoch": 2.320830542531815, "grad_norm": 2.5768882237189827, "learning_rate": 2.2374191999853333e-06, "loss": 1.1181, "step": 3465 }, { "epoch": 2.3215003348961822, "grad_norm": 2.394548697759385, "learning_rate": 2.2359658508609737e-06, "loss": 1.246, "step": 3466 }, { "epoch": 2.322170127260549, "grad_norm": 2.43134248734497, "learning_rate": 2.2345125919694288e-06, "loss": 1.2283, "step": 3467 }, { "epoch": 2.3228399196249163, "grad_norm": 2.4122644792069123, "learning_rate": 2.233059423807347e-06, "loss": 1.0524, "step": 3468 }, { "epoch": 2.323509711989283, "grad_norm": 3.237861927447925, "learning_rate": 2.231606346871343e-06, "loss": 0.7973, "step": 3469 }, { "epoch": 2.3241795043536504, "grad_norm": 2.4102411665112826, "learning_rate": 2.2301533616580006e-06, "loss": 0.9498, "step": 3470 }, { "epoch": 2.3248492967180177, "grad_norm": 2.5408380968435464, "learning_rate": 2.2287004686638726e-06, "loss": 1.2397, "step": 3471 }, { "epoch": 2.3255190890823845, "grad_norm": 2.770256834283307, "learning_rate": 2.227247668385482e-06, "loss": 1.0611, "step": 3472 }, { "epoch": 2.3261888814467513, "grad_norm": 2.3821391733712898, "learning_rate": 2.2257949613193176e-06, "loss": 1.1204, "step": 3473 }, { "epoch": 2.3268586738111185, "grad_norm": 2.6311329094064377, "learning_rate": 2.224342347961837e-06, "loss": 1.0768, "step": 3474 }, { "epoch": 2.327528466175486, "grad_norm": 3.622063895430812, "learning_rate": 2.2228898288094665e-06, "loss": 0.9916, "step": 3475 }, { "epoch": 2.3281982585398526, "grad_norm": 2.895638021095991, "learning_rate": 2.2214374043585996e-06, "loss": 1.0116, "step": 3476 }, { "epoch": 2.3288680509042194, "grad_norm": 2.4752945987205317, "learning_rate": 2.219985075105599e-06, "loss": 1.1606, "step": 3477 }, { "epoch": 2.3295378432685867, "grad_norm": 2.450982324115303, "learning_rate": 2.218532841546791e-06, "loss": 0.9003, "step": 3478 }, { "epoch": 2.330207635632954, "grad_norm": 2.8404265772633357, "learning_rate": 2.2170807041784735e-06, "loss": 0.9383, "step": 3479 }, { "epoch": 2.330877427997321, "grad_norm": 3.1534788165907095, "learning_rate": 2.2156286634969097e-06, "loss": 1.059, "step": 3480 }, { "epoch": 2.331547220361688, "grad_norm": 2.7751977762455646, "learning_rate": 2.21417671999833e-06, "loss": 1.0722, "step": 3481 }, { "epoch": 2.332217012726055, "grad_norm": 2.331758549796877, "learning_rate": 2.2127248741789296e-06, "loss": 0.9974, "step": 3482 }, { "epoch": 2.332886805090422, "grad_norm": 2.4298455390984626, "learning_rate": 2.211273126534874e-06, "loss": 1.2433, "step": 3483 }, { "epoch": 2.333556597454789, "grad_norm": 2.3100944515265787, "learning_rate": 2.209821477562292e-06, "loss": 1.1401, "step": 3484 }, { "epoch": 2.334226389819156, "grad_norm": 2.360025328234934, "learning_rate": 2.208369927757282e-06, "loss": 1.1539, "step": 3485 }, { "epoch": 2.334896182183523, "grad_norm": 2.4652070837859785, "learning_rate": 2.2069184776159035e-06, "loss": 1.1377, "step": 3486 }, { "epoch": 2.3355659745478903, "grad_norm": 3.4647866898964503, "learning_rate": 2.205467127634187e-06, "loss": 1.2222, "step": 3487 }, { "epoch": 2.336235766912257, "grad_norm": 2.3748026674724514, "learning_rate": 2.2040158783081254e-06, "loss": 1.1137, "step": 3488 }, { "epoch": 2.3369055592766244, "grad_norm": 2.5033495682974087, "learning_rate": 2.202564730133681e-06, "loss": 0.9212, "step": 3489 }, { "epoch": 2.337575351640991, "grad_norm": 3.1488962349863328, "learning_rate": 2.2011136836067748e-06, "loss": 0.9748, "step": 3490 }, { "epoch": 2.3382451440053584, "grad_norm": 2.6827969657763346, "learning_rate": 2.1996627392233007e-06, "loss": 1.0485, "step": 3491 }, { "epoch": 2.3389149363697253, "grad_norm": 2.435675263537022, "learning_rate": 2.198211897479112e-06, "loss": 1.1074, "step": 3492 }, { "epoch": 2.3395847287340925, "grad_norm": 3.0111880129614597, "learning_rate": 2.1967611588700326e-06, "loss": 1.188, "step": 3493 }, { "epoch": 2.3402545210984593, "grad_norm": 2.4915698250394542, "learning_rate": 2.1953105238918432e-06, "loss": 1.1798, "step": 3494 }, { "epoch": 2.3409243134628266, "grad_norm": 2.6080858677856233, "learning_rate": 2.1938599930402966e-06, "loss": 1.2297, "step": 3495 }, { "epoch": 2.3415941058271934, "grad_norm": 2.4710295958834494, "learning_rate": 2.1924095668111058e-06, "loss": 1.159, "step": 3496 }, { "epoch": 2.3422638981915607, "grad_norm": 2.3809478105542983, "learning_rate": 2.1909592456999505e-06, "loss": 1.0848, "step": 3497 }, { "epoch": 2.3429336905559275, "grad_norm": 2.265639036860725, "learning_rate": 2.1895090302024717e-06, "loss": 1.1584, "step": 3498 }, { "epoch": 2.3436034829202947, "grad_norm": 2.4822837287986523, "learning_rate": 2.188058920814277e-06, "loss": 1.2432, "step": 3499 }, { "epoch": 2.3442732752846616, "grad_norm": 2.6837093487198764, "learning_rate": 2.186608918030935e-06, "loss": 1.1888, "step": 3500 }, { "epoch": 2.344943067649029, "grad_norm": 2.886635164134186, "learning_rate": 2.1851590223479815e-06, "loss": 1.0654, "step": 3501 }, { "epoch": 2.3456128600133956, "grad_norm": 2.8286344178797984, "learning_rate": 2.1837092342609134e-06, "loss": 1.1901, "step": 3502 }, { "epoch": 2.346282652377763, "grad_norm": 3.2792476636464096, "learning_rate": 2.18225955426519e-06, "loss": 1.0053, "step": 3503 }, { "epoch": 2.34695244474213, "grad_norm": 3.87208473405266, "learning_rate": 2.180809982856234e-06, "loss": 1.0722, "step": 3504 }, { "epoch": 2.347622237106497, "grad_norm": 2.454345277420469, "learning_rate": 2.179360520529434e-06, "loss": 1.0826, "step": 3505 }, { "epoch": 2.348292029470864, "grad_norm": 2.8137574357554938, "learning_rate": 2.177911167780139e-06, "loss": 1.1947, "step": 3506 }, { "epoch": 2.348961821835231, "grad_norm": 2.884862583732252, "learning_rate": 2.1764619251036583e-06, "loss": 1.0629, "step": 3507 }, { "epoch": 2.3496316141995983, "grad_norm": 3.0641201729241394, "learning_rate": 2.175012792995267e-06, "loss": 0.8493, "step": 3508 }, { "epoch": 2.350301406563965, "grad_norm": 2.4178334234295944, "learning_rate": 2.1735637719502018e-06, "loss": 1.052, "step": 3509 }, { "epoch": 2.3509711989283324, "grad_norm": 2.9679371877836433, "learning_rate": 2.172114862463661e-06, "loss": 1.1579, "step": 3510 }, { "epoch": 2.351640991292699, "grad_norm": 2.5732299624775017, "learning_rate": 2.1706660650308043e-06, "loss": 1.1235, "step": 3511 }, { "epoch": 2.3523107836570665, "grad_norm": 2.5748402018492733, "learning_rate": 2.169217380146753e-06, "loss": 1.1084, "step": 3512 }, { "epoch": 2.3529805760214333, "grad_norm": 2.85587746767547, "learning_rate": 2.1677688083065918e-06, "loss": 1.1224, "step": 3513 }, { "epoch": 2.3536503683858006, "grad_norm": 3.6758475746993637, "learning_rate": 2.166320350005365e-06, "loss": 1.1654, "step": 3514 }, { "epoch": 2.3543201607501674, "grad_norm": 2.7831946746595717, "learning_rate": 2.1648720057380776e-06, "loss": 0.9536, "step": 3515 }, { "epoch": 2.3549899531145346, "grad_norm": 4.7342117996726705, "learning_rate": 2.163423775999697e-06, "loss": 1.1581, "step": 3516 }, { "epoch": 2.3556597454789014, "grad_norm": 2.4557154062142157, "learning_rate": 2.161975661285151e-06, "loss": 1.1014, "step": 3517 }, { "epoch": 2.3563295378432687, "grad_norm": 2.607860997926776, "learning_rate": 2.16052766208933e-06, "loss": 0.9827, "step": 3518 }, { "epoch": 2.3569993302076355, "grad_norm": 2.804826898464626, "learning_rate": 2.1590797789070795e-06, "loss": 1.0964, "step": 3519 }, { "epoch": 2.357669122572003, "grad_norm": 2.136794759182448, "learning_rate": 2.157632012233212e-06, "loss": 0.9611, "step": 3520 }, { "epoch": 2.3583389149363696, "grad_norm": 2.755756735927611, "learning_rate": 2.156184362562495e-06, "loss": 1.131, "step": 3521 }, { "epoch": 2.359008707300737, "grad_norm": 2.381591027881933, "learning_rate": 2.15473683038966e-06, "loss": 1.1663, "step": 3522 }, { "epoch": 2.3596784996651037, "grad_norm": 2.389822998536254, "learning_rate": 2.153289416209394e-06, "loss": 1.1082, "step": 3523 }, { "epoch": 2.360348292029471, "grad_norm": 3.189296937342399, "learning_rate": 2.1518421205163485e-06, "loss": 1.2828, "step": 3524 }, { "epoch": 2.3610180843938378, "grad_norm": 2.334191268971568, "learning_rate": 2.1503949438051304e-06, "loss": 1.1615, "step": 3525 }, { "epoch": 2.361687876758205, "grad_norm": 2.365546269189594, "learning_rate": 2.1489478865703096e-06, "loss": 1.1657, "step": 3526 }, { "epoch": 2.362357669122572, "grad_norm": 3.067103876268847, "learning_rate": 2.1475009493064105e-06, "loss": 1.0755, "step": 3527 }, { "epoch": 2.363027461486939, "grad_norm": 2.892745962090777, "learning_rate": 2.1460541325079213e-06, "loss": 1.2817, "step": 3528 }, { "epoch": 2.363697253851306, "grad_norm": 2.7966396553627138, "learning_rate": 2.1446074366692865e-06, "loss": 1.195, "step": 3529 }, { "epoch": 2.364367046215673, "grad_norm": 2.6306170358827274, "learning_rate": 2.14316086228491e-06, "loss": 0.9697, "step": 3530 }, { "epoch": 2.36503683858004, "grad_norm": 3.8389028548178197, "learning_rate": 2.1417144098491525e-06, "loss": 1.1721, "step": 3531 }, { "epoch": 2.3657066309444073, "grad_norm": 2.345673288347225, "learning_rate": 2.140268079856336e-06, "loss": 1.0064, "step": 3532 }, { "epoch": 2.3663764233087745, "grad_norm": 3.6648397789933105, "learning_rate": 2.138821872800738e-06, "loss": 1.016, "step": 3533 }, { "epoch": 2.3670462156731413, "grad_norm": 2.4696128107151463, "learning_rate": 2.137375789176597e-06, "loss": 1.1247, "step": 3534 }, { "epoch": 2.367716008037508, "grad_norm": 2.727923332279986, "learning_rate": 2.1359298294781046e-06, "loss": 1.0682, "step": 3535 }, { "epoch": 2.3683858004018754, "grad_norm": 2.38683538657493, "learning_rate": 2.1344839941994142e-06, "loss": 0.9892, "step": 3536 }, { "epoch": 2.3690555927662427, "grad_norm": 3.523323222629502, "learning_rate": 2.1330382838346356e-06, "loss": 1.0374, "step": 3537 }, { "epoch": 2.3697253851306095, "grad_norm": 2.2141354199952583, "learning_rate": 2.1315926988778358e-06, "loss": 1.1945, "step": 3538 }, { "epoch": 2.3703951774949767, "grad_norm": 3.372753442215417, "learning_rate": 2.1301472398230387e-06, "loss": 0.9119, "step": 3539 }, { "epoch": 2.3710649698593436, "grad_norm": 3.8139938457167513, "learning_rate": 2.1287019071642247e-06, "loss": 1.0254, "step": 3540 }, { "epoch": 2.371734762223711, "grad_norm": 2.9457423883118286, "learning_rate": 2.127256701395331e-06, "loss": 1.1262, "step": 3541 }, { "epoch": 2.3724045545880776, "grad_norm": 3.027756049537887, "learning_rate": 2.1258116230102533e-06, "loss": 1.1574, "step": 3542 }, { "epoch": 2.373074346952445, "grad_norm": 4.194965029639806, "learning_rate": 2.124366672502842e-06, "loss": 1.0328, "step": 3543 }, { "epoch": 2.3737441393168117, "grad_norm": 2.6176334277075126, "learning_rate": 2.1229218503669045e-06, "loss": 1.1391, "step": 3544 }, { "epoch": 2.374413931681179, "grad_norm": 2.265942796000097, "learning_rate": 2.121477157096203e-06, "loss": 1.1025, "step": 3545 }, { "epoch": 2.375083724045546, "grad_norm": 2.2776508716843495, "learning_rate": 2.120032593184458e-06, "loss": 1.0763, "step": 3546 }, { "epoch": 2.375753516409913, "grad_norm": 2.3934357250508715, "learning_rate": 2.118588159125345e-06, "loss": 1.1736, "step": 3547 }, { "epoch": 2.37642330877428, "grad_norm": 2.764568234578234, "learning_rate": 2.1171438554124927e-06, "loss": 1.0145, "step": 3548 }, { "epoch": 2.377093101138647, "grad_norm": 2.3603031879813314, "learning_rate": 2.115699682539488e-06, "loss": 0.9638, "step": 3549 }, { "epoch": 2.377762893503014, "grad_norm": 2.6459378431866973, "learning_rate": 2.1142556409998727e-06, "loss": 1.1587, "step": 3550 }, { "epoch": 2.378432685867381, "grad_norm": 3.6497331407701004, "learning_rate": 2.112811731287144e-06, "loss": 0.9635, "step": 3551 }, { "epoch": 2.379102478231748, "grad_norm": 2.45842402240375, "learning_rate": 2.111367953894752e-06, "loss": 1.1652, "step": 3552 }, { "epoch": 2.3797722705961153, "grad_norm": 2.7729177295557563, "learning_rate": 2.109924309316103e-06, "loss": 0.9855, "step": 3553 }, { "epoch": 2.380442062960482, "grad_norm": 2.4106923404130303, "learning_rate": 2.108480798044559e-06, "loss": 1.1036, "step": 3554 }, { "epoch": 2.3811118553248494, "grad_norm": 2.2980297430654058, "learning_rate": 2.1070374205734346e-06, "loss": 1.1855, "step": 3555 }, { "epoch": 2.381781647689216, "grad_norm": 2.2832029614927283, "learning_rate": 2.105594177396e-06, "loss": 1.0056, "step": 3556 }, { "epoch": 2.3824514400535834, "grad_norm": 2.5431184587144107, "learning_rate": 2.104151069005477e-06, "loss": 1.1527, "step": 3557 }, { "epoch": 2.3831212324179503, "grad_norm": 2.5298262460108214, "learning_rate": 2.1027080958950454e-06, "loss": 1.0604, "step": 3558 }, { "epoch": 2.3837910247823175, "grad_norm": 4.66128186103537, "learning_rate": 2.1012652585578356e-06, "loss": 0.9857, "step": 3559 }, { "epoch": 2.3844608171466843, "grad_norm": 2.4479512206946623, "learning_rate": 2.099822557486933e-06, "loss": 1.2122, "step": 3560 }, { "epoch": 2.3851306095110516, "grad_norm": 2.4687429986823655, "learning_rate": 2.0983799931753746e-06, "loss": 1.2396, "step": 3561 }, { "epoch": 2.385800401875419, "grad_norm": 3.685420841765958, "learning_rate": 2.0969375661161544e-06, "loss": 0.9884, "step": 3562 }, { "epoch": 2.3864701942397857, "grad_norm": 2.3598967248427187, "learning_rate": 2.095495276802216e-06, "loss": 0.9905, "step": 3563 }, { "epoch": 2.3871399866041525, "grad_norm": 2.5693538619695775, "learning_rate": 2.0940531257264562e-06, "loss": 1.0913, "step": 3564 }, { "epoch": 2.3878097789685198, "grad_norm": 2.588506681893894, "learning_rate": 2.092611113381727e-06, "loss": 1.0723, "step": 3565 }, { "epoch": 2.388479571332887, "grad_norm": 2.7683861066638946, "learning_rate": 2.091169240260831e-06, "loss": 1.1552, "step": 3566 }, { "epoch": 2.389149363697254, "grad_norm": 3.4393197854162403, "learning_rate": 2.0897275068565237e-06, "loss": 0.7978, "step": 3567 }, { "epoch": 2.3898191560616207, "grad_norm": 2.476732387582196, "learning_rate": 2.0882859136615116e-06, "loss": 1.1886, "step": 3568 }, { "epoch": 2.390488948425988, "grad_norm": 2.8788551527387476, "learning_rate": 2.086844461168456e-06, "loss": 1.1723, "step": 3569 }, { "epoch": 2.391158740790355, "grad_norm": 2.711485863209293, "learning_rate": 2.0854031498699683e-06, "loss": 0.9961, "step": 3570 }, { "epoch": 2.391828533154722, "grad_norm": 2.3830828069572494, "learning_rate": 2.0839619802586123e-06, "loss": 0.9104, "step": 3571 }, { "epoch": 2.3924983255190893, "grad_norm": 2.5925733614672057, "learning_rate": 2.0825209528269013e-06, "loss": 1.2401, "step": 3572 }, { "epoch": 2.393168117883456, "grad_norm": 4.60493026181484, "learning_rate": 2.081080068067304e-06, "loss": 1.1231, "step": 3573 }, { "epoch": 2.3938379102478233, "grad_norm": 2.367660935150523, "learning_rate": 2.0796393264722366e-06, "loss": 0.96, "step": 3574 }, { "epoch": 2.39450770261219, "grad_norm": 2.2430064194288506, "learning_rate": 2.078198728534068e-06, "loss": 1.1463, "step": 3575 }, { "epoch": 2.3951774949765574, "grad_norm": 4.391698783094375, "learning_rate": 2.07675827474512e-06, "loss": 1.0926, "step": 3576 }, { "epoch": 2.3958472873409242, "grad_norm": 3.016858497488286, "learning_rate": 2.0753179655976598e-06, "loss": 1.1008, "step": 3577 }, { "epoch": 2.3965170797052915, "grad_norm": 2.564139100668081, "learning_rate": 2.07387780158391e-06, "loss": 1.0623, "step": 3578 }, { "epoch": 2.3971868720696583, "grad_norm": 2.528062742981043, "learning_rate": 2.072437783196042e-06, "loss": 1.1081, "step": 3579 }, { "epoch": 2.3978566644340256, "grad_norm": 2.275056606852147, "learning_rate": 2.070997910926178e-06, "loss": 0.9763, "step": 3580 }, { "epoch": 2.3985264567983924, "grad_norm": 2.3290774371150134, "learning_rate": 2.0695581852663884e-06, "loss": 1.0722, "step": 3581 }, { "epoch": 2.3991962491627596, "grad_norm": 2.4179927297484283, "learning_rate": 2.068118606708695e-06, "loss": 1.1209, "step": 3582 }, { "epoch": 2.3998660415271265, "grad_norm": 2.4417693827375424, "learning_rate": 2.06667917574507e-06, "loss": 1.1516, "step": 3583 }, { "epoch": 2.4005358338914937, "grad_norm": 2.300340828452612, "learning_rate": 2.065239892867434e-06, "loss": 1.1815, "step": 3584 }, { "epoch": 2.4012056262558605, "grad_norm": 3.2149919983112603, "learning_rate": 2.0638007585676565e-06, "loss": 1.1467, "step": 3585 }, { "epoch": 2.401875418620228, "grad_norm": 2.5381037574297323, "learning_rate": 2.062361773337557e-06, "loss": 0.9862, "step": 3586 }, { "epoch": 2.4025452109845946, "grad_norm": 3.388698810074475, "learning_rate": 2.0609229376689054e-06, "loss": 1.1036, "step": 3587 }, { "epoch": 2.403215003348962, "grad_norm": 2.5600485374220456, "learning_rate": 2.0594842520534185e-06, "loss": 1.164, "step": 3588 }, { "epoch": 2.4038847957133287, "grad_norm": 3.39677612103008, "learning_rate": 2.0580457169827618e-06, "loss": 1.2321, "step": 3589 }, { "epoch": 2.404554588077696, "grad_norm": 2.511810500557019, "learning_rate": 2.05660733294855e-06, "loss": 1.1024, "step": 3590 }, { "epoch": 2.4052243804420628, "grad_norm": 2.431933288509537, "learning_rate": 2.0551691004423473e-06, "loss": 1.0546, "step": 3591 }, { "epoch": 2.40589417280643, "grad_norm": 2.6229108037281597, "learning_rate": 2.0537310199556644e-06, "loss": 1.2179, "step": 3592 }, { "epoch": 2.406563965170797, "grad_norm": 2.441437799431185, "learning_rate": 2.0522930919799602e-06, "loss": 1.1145, "step": 3593 }, { "epoch": 2.407233757535164, "grad_norm": 2.5542961831424154, "learning_rate": 2.0508553170066417e-06, "loss": 1.0409, "step": 3594 }, { "epoch": 2.4079035498995314, "grad_norm": 2.963600112265906, "learning_rate": 2.049417695527065e-06, "loss": 0.961, "step": 3595 }, { "epoch": 2.408573342263898, "grad_norm": 2.4470562542066188, "learning_rate": 2.047980228032533e-06, "loss": 0.991, "step": 3596 }, { "epoch": 2.409243134628265, "grad_norm": 2.5908074145631486, "learning_rate": 2.0465429150142934e-06, "loss": 1.1561, "step": 3597 }, { "epoch": 2.4099129269926323, "grad_norm": 3.1019203391039127, "learning_rate": 2.0451057569635443e-06, "loss": 1.2633, "step": 3598 }, { "epoch": 2.4105827193569995, "grad_norm": 2.637738642770013, "learning_rate": 2.0436687543714303e-06, "loss": 0.9542, "step": 3599 }, { "epoch": 2.4112525117213663, "grad_norm": 2.523440001565255, "learning_rate": 2.0422319077290424e-06, "loss": 1.2169, "step": 3600 }, { "epoch": 2.4119223040857336, "grad_norm": 2.4175143025795824, "learning_rate": 2.0407952175274177e-06, "loss": 0.9345, "step": 3601 }, { "epoch": 2.4125920964501004, "grad_norm": 2.352648343828003, "learning_rate": 2.0393586842575403e-06, "loss": 0.9513, "step": 3602 }, { "epoch": 2.4132618888144677, "grad_norm": 2.614708348562329, "learning_rate": 2.0379223084103413e-06, "loss": 1.1979, "step": 3603 }, { "epoch": 2.4139316811788345, "grad_norm": 2.5574697001230904, "learning_rate": 2.036486090476698e-06, "loss": 1.17, "step": 3604 }, { "epoch": 2.4146014735432018, "grad_norm": 2.7206115335846164, "learning_rate": 2.0350500309474326e-06, "loss": 1.1065, "step": 3605 }, { "epoch": 2.4152712659075686, "grad_norm": 2.735060114637873, "learning_rate": 2.0336141303133125e-06, "loss": 1.2313, "step": 3606 }, { "epoch": 2.415941058271936, "grad_norm": 3.195955783269673, "learning_rate": 2.0321783890650543e-06, "loss": 1.0764, "step": 3607 }, { "epoch": 2.4166108506363027, "grad_norm": 3.4471757801393323, "learning_rate": 2.0307428076933162e-06, "loss": 1.1613, "step": 3608 }, { "epoch": 2.41728064300067, "grad_norm": 2.4334964785423243, "learning_rate": 2.0293073866887057e-06, "loss": 1.1649, "step": 3609 }, { "epoch": 2.4179504353650367, "grad_norm": 2.3377855244636883, "learning_rate": 2.027872126541771e-06, "loss": 1.1471, "step": 3610 }, { "epoch": 2.418620227729404, "grad_norm": 2.8519699737803634, "learning_rate": 2.0264370277430083e-06, "loss": 1.0038, "step": 3611 }, { "epoch": 2.419290020093771, "grad_norm": 2.578177761318339, "learning_rate": 2.025002090782858e-06, "loss": 1.2177, "step": 3612 }, { "epoch": 2.419959812458138, "grad_norm": 2.3751446732674304, "learning_rate": 2.0235673161517066e-06, "loss": 1.2524, "step": 3613 }, { "epoch": 2.420629604822505, "grad_norm": 2.4640815820325512, "learning_rate": 2.0221327043398814e-06, "loss": 1.168, "step": 3614 }, { "epoch": 2.421299397186872, "grad_norm": 3.8436686519833763, "learning_rate": 2.0206982558376573e-06, "loss": 1.1378, "step": 3615 }, { "epoch": 2.421969189551239, "grad_norm": 2.645301938832366, "learning_rate": 2.0192639711352523e-06, "loss": 0.9539, "step": 3616 }, { "epoch": 2.4226389819156062, "grad_norm": 2.5067683781905132, "learning_rate": 2.0178298507228296e-06, "loss": 1.1671, "step": 3617 }, { "epoch": 2.423308774279973, "grad_norm": 2.4762364785283726, "learning_rate": 2.016395895090493e-06, "loss": 1.0575, "step": 3618 }, { "epoch": 2.4239785666443403, "grad_norm": 2.3372785681050154, "learning_rate": 2.0149621047282945e-06, "loss": 1.0593, "step": 3619 }, { "epoch": 2.424648359008707, "grad_norm": 2.444511770466335, "learning_rate": 2.0135284801262255e-06, "loss": 1.0303, "step": 3620 }, { "epoch": 2.4253181513730744, "grad_norm": 3.718632145045211, "learning_rate": 2.0120950217742248e-06, "loss": 0.9794, "step": 3621 }, { "epoch": 2.425987943737441, "grad_norm": 4.018057287610034, "learning_rate": 2.01066173016217e-06, "loss": 0.9315, "step": 3622 }, { "epoch": 2.4266577361018085, "grad_norm": 2.3495279199812478, "learning_rate": 2.0092286057798836e-06, "loss": 0.9854, "step": 3623 }, { "epoch": 2.4273275284661757, "grad_norm": 2.3738361096586558, "learning_rate": 2.007795649117133e-06, "loss": 1.0168, "step": 3624 }, { "epoch": 2.4279973208305425, "grad_norm": 2.991222246413617, "learning_rate": 2.0063628606636265e-06, "loss": 0.9795, "step": 3625 }, { "epoch": 2.4286671131949094, "grad_norm": 2.7764655807935665, "learning_rate": 2.0049302409090134e-06, "loss": 1.0217, "step": 3626 }, { "epoch": 2.4293369055592766, "grad_norm": 2.412807108774105, "learning_rate": 2.003497790342887e-06, "loss": 0.9972, "step": 3627 }, { "epoch": 2.430006697923644, "grad_norm": 2.712305428608225, "learning_rate": 2.0020655094547843e-06, "loss": 1.1244, "step": 3628 }, { "epoch": 2.4306764902880107, "grad_norm": 3.6873592085223024, "learning_rate": 2.0006333987341817e-06, "loss": 1.1417, "step": 3629 }, { "epoch": 2.431346282652378, "grad_norm": 4.617237791417098, "learning_rate": 1.9992014586704976e-06, "loss": 1.0101, "step": 3630 }, { "epoch": 2.4320160750167448, "grad_norm": 2.6356518446277946, "learning_rate": 1.9977696897530934e-06, "loss": 1.2355, "step": 3631 }, { "epoch": 2.432685867381112, "grad_norm": 3.0805849064918527, "learning_rate": 1.996338092471272e-06, "loss": 1.0883, "step": 3632 }, { "epoch": 2.433355659745479, "grad_norm": 2.2370262815665476, "learning_rate": 1.994906667314278e-06, "loss": 0.9878, "step": 3633 }, { "epoch": 2.434025452109846, "grad_norm": 2.8288410601013343, "learning_rate": 1.993475414771294e-06, "loss": 0.921, "step": 3634 }, { "epoch": 2.434695244474213, "grad_norm": 2.5196265165824454, "learning_rate": 1.9920443353314463e-06, "loss": 0.9956, "step": 3635 }, { "epoch": 2.43536503683858, "grad_norm": 2.312702565952905, "learning_rate": 1.9906134294838036e-06, "loss": 1.0271, "step": 3636 }, { "epoch": 2.436034829202947, "grad_norm": 2.3389682041498556, "learning_rate": 1.989182697717372e-06, "loss": 1.088, "step": 3637 }, { "epoch": 2.4367046215673143, "grad_norm": 2.5859065135066617, "learning_rate": 1.9877521405210996e-06, "loss": 1.1653, "step": 3638 }, { "epoch": 2.437374413931681, "grad_norm": 2.417414069066383, "learning_rate": 1.9863217583838735e-06, "loss": 0.9877, "step": 3639 }, { "epoch": 2.4380442062960483, "grad_norm": 2.5764709798823913, "learning_rate": 1.9848915517945247e-06, "loss": 1.1382, "step": 3640 }, { "epoch": 2.438713998660415, "grad_norm": 2.5427273455813073, "learning_rate": 1.9834615212418205e-06, "loss": 1.0362, "step": 3641 }, { "epoch": 2.4393837910247824, "grad_norm": 2.800916527755613, "learning_rate": 1.9820316672144684e-06, "loss": 0.9531, "step": 3642 }, { "epoch": 2.4400535833891492, "grad_norm": 2.495654994302732, "learning_rate": 1.980601990201117e-06, "loss": 1.0641, "step": 3643 }, { "epoch": 2.4407233757535165, "grad_norm": 2.3932013273342854, "learning_rate": 1.979172490690354e-06, "loss": 1.016, "step": 3644 }, { "epoch": 2.4413931681178833, "grad_norm": 2.4714940539116372, "learning_rate": 1.9777431691707054e-06, "loss": 1.0982, "step": 3645 }, { "epoch": 2.4420629604822506, "grad_norm": 2.482008883856617, "learning_rate": 1.9763140261306395e-06, "loss": 1.1354, "step": 3646 }, { "epoch": 2.4427327528466174, "grad_norm": 2.8535151354624304, "learning_rate": 1.974885062058558e-06, "loss": 1.2405, "step": 3647 }, { "epoch": 2.4434025452109847, "grad_norm": 2.818749376461753, "learning_rate": 1.973456277442807e-06, "loss": 1.1437, "step": 3648 }, { "epoch": 2.4440723375753515, "grad_norm": 2.546265932841901, "learning_rate": 1.9720276727716677e-06, "loss": 1.0655, "step": 3649 }, { "epoch": 2.4447421299397187, "grad_norm": 3.229032818236825, "learning_rate": 1.970599248533364e-06, "loss": 1.048, "step": 3650 }, { "epoch": 2.4454119223040856, "grad_norm": 2.4477727548960506, "learning_rate": 1.969171005216051e-06, "loss": 1.005, "step": 3651 }, { "epoch": 2.446081714668453, "grad_norm": 2.475440986864812, "learning_rate": 1.967742943307829e-06, "loss": 1.1154, "step": 3652 }, { "epoch": 2.44675150703282, "grad_norm": 2.3935394522734903, "learning_rate": 1.9663150632967324e-06, "loss": 1.0253, "step": 3653 }, { "epoch": 2.447421299397187, "grad_norm": 3.0216495754269155, "learning_rate": 1.964887365670736e-06, "loss": 1.016, "step": 3654 }, { "epoch": 2.4480910917615537, "grad_norm": 2.34515888185374, "learning_rate": 1.9634598509177494e-06, "loss": 1.0651, "step": 3655 }, { "epoch": 2.448760884125921, "grad_norm": 3.6135845887906313, "learning_rate": 1.9620325195256216e-06, "loss": 1.1935, "step": 3656 }, { "epoch": 2.4494306764902882, "grad_norm": 2.690676015742516, "learning_rate": 1.960605371982138e-06, "loss": 0.9866, "step": 3657 }, { "epoch": 2.450100468854655, "grad_norm": 4.0499283848856305, "learning_rate": 1.959178408775023e-06, "loss": 1.2675, "step": 3658 }, { "epoch": 2.450770261219022, "grad_norm": 2.3083447274292, "learning_rate": 1.9577516303919355e-06, "loss": 1.046, "step": 3659 }, { "epoch": 2.451440053583389, "grad_norm": 2.268573159782296, "learning_rate": 1.9563250373204727e-06, "loss": 1.016, "step": 3660 }, { "epoch": 2.4521098459477564, "grad_norm": 2.545813795650038, "learning_rate": 1.9548986300481674e-06, "loss": 1.1242, "step": 3661 }, { "epoch": 2.452779638312123, "grad_norm": 2.4758472792078328, "learning_rate": 1.953472409062492e-06, "loss": 1.1525, "step": 3662 }, { "epoch": 2.4534494306764905, "grad_norm": 3.2219563575434154, "learning_rate": 1.9520463748508505e-06, "loss": 1.1779, "step": 3663 }, { "epoch": 2.4541192230408573, "grad_norm": 2.9115690536505214, "learning_rate": 1.9506205279005862e-06, "loss": 1.0519, "step": 3664 }, { "epoch": 2.4547890154052245, "grad_norm": 2.4186076197663002, "learning_rate": 1.9491948686989774e-06, "loss": 1.1647, "step": 3665 }, { "epoch": 2.4554588077695914, "grad_norm": 2.3425214679793807, "learning_rate": 1.94776939773324e-06, "loss": 0.9637, "step": 3666 }, { "epoch": 2.4561286001339586, "grad_norm": 3.6349341992379864, "learning_rate": 1.9463441154905222e-06, "loss": 1.0935, "step": 3667 }, { "epoch": 2.4567983924983254, "grad_norm": 2.7887879160674602, "learning_rate": 1.9449190224579102e-06, "loss": 1.0919, "step": 3668 }, { "epoch": 2.4574681848626927, "grad_norm": 3.303035334074844, "learning_rate": 1.9434941191224255e-06, "loss": 0.9778, "step": 3669 }, { "epoch": 2.4581379772270595, "grad_norm": 2.6656459748003734, "learning_rate": 1.9420694059710245e-06, "loss": 0.9288, "step": 3670 }, { "epoch": 2.4588077695914268, "grad_norm": 2.5291625966847637, "learning_rate": 1.940644883490597e-06, "loss": 1.217, "step": 3671 }, { "epoch": 2.4594775619557936, "grad_norm": 2.5807148596344014, "learning_rate": 1.9392205521679688e-06, "loss": 1.1145, "step": 3672 }, { "epoch": 2.460147354320161, "grad_norm": 3.049920151341528, "learning_rate": 1.9377964124899023e-06, "loss": 1.089, "step": 3673 }, { "epoch": 2.4608171466845277, "grad_norm": 2.491921044301245, "learning_rate": 1.936372464943092e-06, "loss": 1.0733, "step": 3674 }, { "epoch": 2.461486939048895, "grad_norm": 2.5809223654135622, "learning_rate": 1.934948710014166e-06, "loss": 1.0913, "step": 3675 }, { "epoch": 2.4621567314132617, "grad_norm": 2.930540857603616, "learning_rate": 1.9335251481896887e-06, "loss": 1.0765, "step": 3676 }, { "epoch": 2.462826523777629, "grad_norm": 2.815768212540405, "learning_rate": 1.932101779956158e-06, "loss": 0.9556, "step": 3677 }, { "epoch": 2.463496316141996, "grad_norm": 2.3953594652368477, "learning_rate": 1.9306786058000067e-06, "loss": 0.9681, "step": 3678 }, { "epoch": 2.464166108506363, "grad_norm": 3.035947438871125, "learning_rate": 1.9292556262075977e-06, "loss": 1.0965, "step": 3679 }, { "epoch": 2.46483590087073, "grad_norm": 2.706750314013715, "learning_rate": 1.9278328416652305e-06, "loss": 1.1071, "step": 3680 }, { "epoch": 2.465505693235097, "grad_norm": 3.0051215010844516, "learning_rate": 1.9264102526591367e-06, "loss": 1.0008, "step": 3681 }, { "epoch": 2.466175485599464, "grad_norm": 3.2317322638263533, "learning_rate": 1.9249878596754824e-06, "loss": 0.9276, "step": 3682 }, { "epoch": 2.4668452779638312, "grad_norm": 3.0747274912323395, "learning_rate": 1.9235656632003662e-06, "loss": 1.0599, "step": 3683 }, { "epoch": 2.467515070328198, "grad_norm": 2.4391038551299884, "learning_rate": 1.9221436637198174e-06, "loss": 1.0932, "step": 3684 }, { "epoch": 2.4681848626925653, "grad_norm": 3.240803564170188, "learning_rate": 1.9207218617198006e-06, "loss": 1.1299, "step": 3685 }, { "epoch": 2.4688546550569326, "grad_norm": 2.446002520873495, "learning_rate": 1.919300257686212e-06, "loss": 1.0262, "step": 3686 }, { "epoch": 2.4695244474212994, "grad_norm": 3.101704529572967, "learning_rate": 1.917878852104881e-06, "loss": 1.0894, "step": 3687 }, { "epoch": 2.470194239785666, "grad_norm": 2.633476517417037, "learning_rate": 1.916457645461567e-06, "loss": 1.1399, "step": 3688 }, { "epoch": 2.4708640321500335, "grad_norm": 2.7072678806939865, "learning_rate": 1.9150366382419634e-06, "loss": 1.0607, "step": 3689 }, { "epoch": 2.4715338245144007, "grad_norm": 2.6819969238499084, "learning_rate": 1.9136158309316945e-06, "loss": 1.2054, "step": 3690 }, { "epoch": 2.4722036168787676, "grad_norm": 3.2149511052692765, "learning_rate": 1.9121952240163182e-06, "loss": 1.0363, "step": 3691 }, { "epoch": 2.472873409243135, "grad_norm": 2.6621899486237384, "learning_rate": 1.9107748179813194e-06, "loss": 1.0696, "step": 3692 }, { "epoch": 2.4735432016075016, "grad_norm": 2.40224089316365, "learning_rate": 1.9093546133121197e-06, "loss": 1.0363, "step": 3693 }, { "epoch": 2.474212993971869, "grad_norm": 9.286470117988253, "learning_rate": 1.9079346104940675e-06, "loss": 1.1028, "step": 3694 }, { "epoch": 2.4748827863362357, "grad_norm": 3.0226505223827114, "learning_rate": 1.906514810012447e-06, "loss": 1.1251, "step": 3695 }, { "epoch": 2.475552578700603, "grad_norm": 2.4680457874713704, "learning_rate": 1.9050952123524677e-06, "loss": 1.1502, "step": 3696 }, { "epoch": 2.47622237106497, "grad_norm": 2.473470190667551, "learning_rate": 1.9036758179992735e-06, "loss": 1.185, "step": 3697 }, { "epoch": 2.476892163429337, "grad_norm": 2.4612980333054613, "learning_rate": 1.9022566274379373e-06, "loss": 1.1679, "step": 3698 }, { "epoch": 2.477561955793704, "grad_norm": 3.1621033010891644, "learning_rate": 1.9008376411534645e-06, "loss": 0.9727, "step": 3699 }, { "epoch": 2.478231748158071, "grad_norm": 2.536424854084009, "learning_rate": 1.8994188596307874e-06, "loss": 1.1576, "step": 3700 }, { "epoch": 2.478901540522438, "grad_norm": 2.4478128148928886, "learning_rate": 1.8980002833547706e-06, "loss": 1.0013, "step": 3701 }, { "epoch": 2.479571332886805, "grad_norm": 2.333938575033503, "learning_rate": 1.8965819128102076e-06, "loss": 1.1186, "step": 3702 }, { "epoch": 2.480241125251172, "grad_norm": 3.1598437774110546, "learning_rate": 1.8951637484818227e-06, "loss": 0.9672, "step": 3703 }, { "epoch": 2.4809109176155393, "grad_norm": 2.4401800999418946, "learning_rate": 1.8937457908542684e-06, "loss": 1.0524, "step": 3704 }, { "epoch": 2.481580709979906, "grad_norm": 2.6763370476033312, "learning_rate": 1.8923280404121272e-06, "loss": 1.1712, "step": 3705 }, { "epoch": 2.4822505023442734, "grad_norm": 2.5633151596892167, "learning_rate": 1.8909104976399101e-06, "loss": 1.0653, "step": 3706 }, { "epoch": 2.48292029470864, "grad_norm": 2.651545133546354, "learning_rate": 1.8894931630220599e-06, "loss": 1.071, "step": 3707 }, { "epoch": 2.4835900870730074, "grad_norm": 3.3976484802249525, "learning_rate": 1.8880760370429435e-06, "loss": 0.9277, "step": 3708 }, { "epoch": 2.4842598794373743, "grad_norm": 2.5085054307367267, "learning_rate": 1.8866591201868606e-06, "loss": 1.0886, "step": 3709 }, { "epoch": 2.4849296718017415, "grad_norm": 2.585939496151858, "learning_rate": 1.885242412938037e-06, "loss": 1.1336, "step": 3710 }, { "epoch": 2.4855994641661083, "grad_norm": 2.387320770988593, "learning_rate": 1.8838259157806293e-06, "loss": 1.0158, "step": 3711 }, { "epoch": 2.4862692565304756, "grad_norm": 2.563275132893803, "learning_rate": 1.8824096291987195e-06, "loss": 1.1153, "step": 3712 }, { "epoch": 2.4869390488948424, "grad_norm": 2.626730255630526, "learning_rate": 1.8809935536763188e-06, "loss": 0.9412, "step": 3713 }, { "epoch": 2.4876088412592097, "grad_norm": 3.617944333483685, "learning_rate": 1.8795776896973678e-06, "loss": 1.1383, "step": 3714 }, { "epoch": 2.488278633623577, "grad_norm": 2.813840599577998, "learning_rate": 1.8781620377457322e-06, "loss": 1.2344, "step": 3715 }, { "epoch": 2.4889484259879437, "grad_norm": 3.986251717771414, "learning_rate": 1.8767465983052078e-06, "loss": 1.1822, "step": 3716 }, { "epoch": 2.4896182183523106, "grad_norm": 2.6411532988524793, "learning_rate": 1.8753313718595142e-06, "loss": 1.0976, "step": 3717 }, { "epoch": 2.490288010716678, "grad_norm": 2.4217452144226277, "learning_rate": 1.8739163588923026e-06, "loss": 1.1121, "step": 3718 }, { "epoch": 2.490957803081045, "grad_norm": 2.5892420538534346, "learning_rate": 1.8725015598871481e-06, "loss": 1.1807, "step": 3719 }, { "epoch": 2.491627595445412, "grad_norm": 2.7398429568482214, "learning_rate": 1.8710869753275545e-06, "loss": 1.1439, "step": 3720 }, { "epoch": 2.4922973878097787, "grad_norm": 2.4604191983995713, "learning_rate": 1.86967260569695e-06, "loss": 1.1985, "step": 3721 }, { "epoch": 2.492967180174146, "grad_norm": 2.432051859607002, "learning_rate": 1.8682584514786923e-06, "loss": 1.1358, "step": 3722 }, { "epoch": 2.4936369725385132, "grad_norm": 3.4127282842597295, "learning_rate": 1.8668445131560631e-06, "loss": 0.965, "step": 3723 }, { "epoch": 2.49430676490288, "grad_norm": 2.8796131693540183, "learning_rate": 1.8654307912122721e-06, "loss": 1.073, "step": 3724 }, { "epoch": 2.4949765572672473, "grad_norm": 2.855915019548487, "learning_rate": 1.864017286130453e-06, "loss": 1.214, "step": 3725 }, { "epoch": 2.495646349631614, "grad_norm": 3.9274624279980057, "learning_rate": 1.8626039983936675e-06, "loss": 0.7165, "step": 3726 }, { "epoch": 2.4963161419959814, "grad_norm": 2.3694633197037764, "learning_rate": 1.8611909284849018e-06, "loss": 0.9782, "step": 3727 }, { "epoch": 2.496985934360348, "grad_norm": 3.851080035427194, "learning_rate": 1.859778076887069e-06, "loss": 1.1192, "step": 3728 }, { "epoch": 2.4976557267247155, "grad_norm": 2.8403607063773153, "learning_rate": 1.8583654440830042e-06, "loss": 1.156, "step": 3729 }, { "epoch": 2.4983255190890823, "grad_norm": 2.5424676975222837, "learning_rate": 1.8569530305554722e-06, "loss": 1.1995, "step": 3730 }, { "epoch": 2.4989953114534496, "grad_norm": 3.208667121866662, "learning_rate": 1.8555408367871598e-06, "loss": 1.1343, "step": 3731 }, { "epoch": 2.4996651038178164, "grad_norm": 2.6934033194389113, "learning_rate": 1.8541288632606807e-06, "loss": 0.7597, "step": 3732 }, { "epoch": 2.5003348961821836, "grad_norm": 3.6093425085530937, "learning_rate": 1.8527171104585706e-06, "loss": 0.887, "step": 3733 }, { "epoch": 2.5010046885465504, "grad_norm": 2.476784579111272, "learning_rate": 1.8513055788632925e-06, "loss": 1.1474, "step": 3734 }, { "epoch": 2.5016744809109177, "grad_norm": 2.515423859697332, "learning_rate": 1.849894268957232e-06, "loss": 1.0123, "step": 3735 }, { "epoch": 2.5023442732752845, "grad_norm": 3.1781092041076815, "learning_rate": 1.8484831812227022e-06, "loss": 1.0967, "step": 3736 }, { "epoch": 2.503014065639652, "grad_norm": 4.48679195447411, "learning_rate": 1.8470723161419342e-06, "loss": 0.8334, "step": 3737 }, { "epoch": 2.5036838580040186, "grad_norm": 3.3779875734471068, "learning_rate": 1.8456616741970883e-06, "loss": 1.0747, "step": 3738 }, { "epoch": 2.504353650368386, "grad_norm": 2.740999248146935, "learning_rate": 1.8442512558702463e-06, "loss": 1.1016, "step": 3739 }, { "epoch": 2.5050234427327527, "grad_norm": 2.232898198506164, "learning_rate": 1.842841061643416e-06, "loss": 0.9781, "step": 3740 }, { "epoch": 2.50569323509712, "grad_norm": 2.8366297564148764, "learning_rate": 1.8414310919985242e-06, "loss": 0.9407, "step": 3741 }, { "epoch": 2.5063630274614868, "grad_norm": 2.354347187336083, "learning_rate": 1.8400213474174248e-06, "loss": 1.047, "step": 3742 }, { "epoch": 2.507032819825854, "grad_norm": 2.9416841253051, "learning_rate": 1.8386118283818926e-06, "loss": 0.9168, "step": 3743 }, { "epoch": 2.5077026121902213, "grad_norm": 2.718292285635765, "learning_rate": 1.8372025353736279e-06, "loss": 1.1577, "step": 3744 }, { "epoch": 2.508372404554588, "grad_norm": 2.372311411011058, "learning_rate": 1.83579346887425e-06, "loss": 1.0073, "step": 3745 }, { "epoch": 2.509042196918955, "grad_norm": 2.138009927645596, "learning_rate": 1.8343846293653033e-06, "loss": 1.0283, "step": 3746 }, { "epoch": 2.509711989283322, "grad_norm": 2.6385137371320884, "learning_rate": 1.8329760173282545e-06, "loss": 1.0406, "step": 3747 }, { "epoch": 2.5103817816476894, "grad_norm": 3.1510034368596074, "learning_rate": 1.831567633244493e-06, "loss": 0.9929, "step": 3748 }, { "epoch": 2.5110515740120563, "grad_norm": 2.7487530216971314, "learning_rate": 1.8301594775953286e-06, "loss": 1.0737, "step": 3749 }, { "epoch": 2.511721366376423, "grad_norm": 2.7221075801344723, "learning_rate": 1.8287515508619934e-06, "loss": 0.7675, "step": 3750 }, { "epoch": 2.5123911587407903, "grad_norm": 2.597375109194706, "learning_rate": 1.8273438535256426e-06, "loss": 1.0575, "step": 3751 }, { "epoch": 2.5130609511051576, "grad_norm": 3.7337771782610525, "learning_rate": 1.825936386067352e-06, "loss": 1.1825, "step": 3752 }, { "epoch": 2.5137307434695244, "grad_norm": 3.1258823363587402, "learning_rate": 1.82452914896812e-06, "loss": 1.0699, "step": 3753 }, { "epoch": 2.5144005358338912, "grad_norm": 4.4568587720436135, "learning_rate": 1.8231221427088635e-06, "loss": 0.9255, "step": 3754 }, { "epoch": 2.5150703281982585, "grad_norm": 2.7150622787027765, "learning_rate": 1.821715367770423e-06, "loss": 1.1719, "step": 3755 }, { "epoch": 2.5157401205626257, "grad_norm": 2.6701231280076825, "learning_rate": 1.8203088246335599e-06, "loss": 1.1969, "step": 3756 }, { "epoch": 2.5164099129269926, "grad_norm": 3.0880349777802745, "learning_rate": 1.8189025137789556e-06, "loss": 1.1153, "step": 3757 }, { "epoch": 2.51707970529136, "grad_norm": 4.293266763674408, "learning_rate": 1.8174964356872116e-06, "loss": 0.9658, "step": 3758 }, { "epoch": 2.5177494976557266, "grad_norm": 2.708325367369476, "learning_rate": 1.816090590838851e-06, "loss": 1.155, "step": 3759 }, { "epoch": 2.518419290020094, "grad_norm": 2.593189906287549, "learning_rate": 1.814684979714317e-06, "loss": 1.1281, "step": 3760 }, { "epoch": 2.5190890823844607, "grad_norm": 2.6523271421113863, "learning_rate": 1.813279602793973e-06, "loss": 1.07, "step": 3761 }, { "epoch": 2.519758874748828, "grad_norm": 2.90966306923845, "learning_rate": 1.8118744605581001e-06, "loss": 1.1777, "step": 3762 }, { "epoch": 2.520428667113195, "grad_norm": 2.8510695422938577, "learning_rate": 1.8104695534869032e-06, "loss": 1.0717, "step": 3763 }, { "epoch": 2.521098459477562, "grad_norm": 2.5897905394624523, "learning_rate": 1.8090648820605036e-06, "loss": 0.981, "step": 3764 }, { "epoch": 2.521768251841929, "grad_norm": 2.5678548827140895, "learning_rate": 1.8076604467589442e-06, "loss": 1.0667, "step": 3765 }, { "epoch": 2.522438044206296, "grad_norm": 2.90835439472202, "learning_rate": 1.8062562480621846e-06, "loss": 1.0179, "step": 3766 }, { "epoch": 2.523107836570663, "grad_norm": 2.7848323951201612, "learning_rate": 1.8048522864501067e-06, "loss": 0.8093, "step": 3767 }, { "epoch": 2.52377762893503, "grad_norm": 3.2129827081633913, "learning_rate": 1.803448562402509e-06, "loss": 1.0609, "step": 3768 }, { "epoch": 2.524447421299397, "grad_norm": 3.5896668424016105, "learning_rate": 1.8020450763991104e-06, "loss": 0.8846, "step": 3769 }, { "epoch": 2.5251172136637643, "grad_norm": 3.192016131962844, "learning_rate": 1.8006418289195465e-06, "loss": 1.1758, "step": 3770 }, { "epoch": 2.525787006028131, "grad_norm": 4.454524477478592, "learning_rate": 1.7992388204433735e-06, "loss": 1.0503, "step": 3771 }, { "epoch": 2.5264567983924984, "grad_norm": 2.7161802283825476, "learning_rate": 1.7978360514500646e-06, "loss": 1.1997, "step": 3772 }, { "epoch": 2.5271265907568656, "grad_norm": 2.968322750050524, "learning_rate": 1.7964335224190133e-06, "loss": 1.0485, "step": 3773 }, { "epoch": 2.5277963831212324, "grad_norm": 3.874961853060294, "learning_rate": 1.795031233829526e-06, "loss": 0.9492, "step": 3774 }, { "epoch": 2.5284661754855993, "grad_norm": 2.473470007874331, "learning_rate": 1.7936291861608324e-06, "loss": 1.1919, "step": 3775 }, { "epoch": 2.5291359678499665, "grad_norm": 2.400459896088305, "learning_rate": 1.7922273798920776e-06, "loss": 1.218, "step": 3776 }, { "epoch": 2.529805760214334, "grad_norm": 3.5857142583524877, "learning_rate": 1.790825815502325e-06, "loss": 0.8969, "step": 3777 }, { "epoch": 2.5304755525787006, "grad_norm": 13.712904506754183, "learning_rate": 1.7894244934705528e-06, "loss": 1.139, "step": 3778 }, { "epoch": 2.5311453449430674, "grad_norm": 2.462138320868334, "learning_rate": 1.7880234142756598e-06, "loss": 1.0455, "step": 3779 }, { "epoch": 2.5318151373074347, "grad_norm": 2.4898043759616093, "learning_rate": 1.7866225783964592e-06, "loss": 0.9463, "step": 3780 }, { "epoch": 2.532484929671802, "grad_norm": 3.0506939341504653, "learning_rate": 1.7852219863116846e-06, "loss": 0.9243, "step": 3781 }, { "epoch": 2.5331547220361688, "grad_norm": 2.4744520046755323, "learning_rate": 1.7838216384999802e-06, "loss": 0.9761, "step": 3782 }, { "epoch": 2.5338245144005356, "grad_norm": 2.8122293319581644, "learning_rate": 1.7824215354399127e-06, "loss": 1.1391, "step": 3783 }, { "epoch": 2.534494306764903, "grad_norm": 3.6428213586690785, "learning_rate": 1.7810216776099615e-06, "loss": 1.2645, "step": 3784 }, { "epoch": 2.53516409912927, "grad_norm": 4.460716655430315, "learning_rate": 1.7796220654885254e-06, "loss": 0.9832, "step": 3785 }, { "epoch": 2.535833891493637, "grad_norm": 2.5899755175075443, "learning_rate": 1.7782226995539155e-06, "loss": 1.0922, "step": 3786 }, { "epoch": 2.536503683858004, "grad_norm": 2.5557136391335185, "learning_rate": 1.7768235802843615e-06, "loss": 1.0849, "step": 3787 }, { "epoch": 2.537173476222371, "grad_norm": 3.843681329103784, "learning_rate": 1.7754247081580066e-06, "loss": 1.095, "step": 3788 }, { "epoch": 2.5378432685867383, "grad_norm": 2.3840796934506137, "learning_rate": 1.7740260836529124e-06, "loss": 1.0313, "step": 3789 }, { "epoch": 2.538513060951105, "grad_norm": 2.705867386367704, "learning_rate": 1.7726277072470543e-06, "loss": 1.0788, "step": 3790 }, { "epoch": 2.5391828533154723, "grad_norm": 2.6164549421276724, "learning_rate": 1.7712295794183215e-06, "loss": 1.0689, "step": 3791 }, { "epoch": 2.539852645679839, "grad_norm": 2.8766098199097643, "learning_rate": 1.76983170064452e-06, "loss": 0.9431, "step": 3792 }, { "epoch": 2.5405224380442064, "grad_norm": 4.203742765226229, "learning_rate": 1.7684340714033716e-06, "loss": 0.9372, "step": 3793 }, { "epoch": 2.5411922304085732, "grad_norm": 3.2637043730708437, "learning_rate": 1.7670366921725106e-06, "loss": 1.2665, "step": 3794 }, { "epoch": 2.5418620227729405, "grad_norm": 2.2910222981872033, "learning_rate": 1.7656395634294871e-06, "loss": 0.8505, "step": 3795 }, { "epoch": 2.5425318151373073, "grad_norm": 2.7876322156063984, "learning_rate": 1.7642426856517645e-06, "loss": 1.2332, "step": 3796 }, { "epoch": 2.5432016075016746, "grad_norm": 2.453251987670233, "learning_rate": 1.7628460593167223e-06, "loss": 1.2157, "step": 3797 }, { "epoch": 2.5438713998660414, "grad_norm": 2.547914995923535, "learning_rate": 1.7614496849016532e-06, "loss": 1.1957, "step": 3798 }, { "epoch": 2.5445411922304086, "grad_norm": 3.8332210745757562, "learning_rate": 1.7600535628837628e-06, "loss": 1.0727, "step": 3799 }, { "epoch": 2.5452109845947755, "grad_norm": 2.454066997499784, "learning_rate": 1.7586576937401712e-06, "loss": 0.9243, "step": 3800 }, { "epoch": 2.5458807769591427, "grad_norm": 2.384772273291479, "learning_rate": 1.7572620779479127e-06, "loss": 0.8959, "step": 3801 }, { "epoch": 2.54655056932351, "grad_norm": 2.5838808948121024, "learning_rate": 1.755866715983935e-06, "loss": 1.0964, "step": 3802 }, { "epoch": 2.547220361687877, "grad_norm": 3.105579583706334, "learning_rate": 1.7544716083250968e-06, "loss": 1.201, "step": 3803 }, { "epoch": 2.5478901540522436, "grad_norm": 3.2670481731756045, "learning_rate": 1.753076755448173e-06, "loss": 1.0, "step": 3804 }, { "epoch": 2.548559946416611, "grad_norm": 2.6686482533356, "learning_rate": 1.75168215782985e-06, "loss": 1.0425, "step": 3805 }, { "epoch": 2.549229738780978, "grad_norm": 2.6119242501781206, "learning_rate": 1.7502878159467274e-06, "loss": 1.1098, "step": 3806 }, { "epoch": 2.549899531145345, "grad_norm": 2.4916195336853635, "learning_rate": 1.7488937302753154e-06, "loss": 1.0686, "step": 3807 }, { "epoch": 2.5505693235097118, "grad_norm": 2.7170638513859773, "learning_rate": 1.7474999012920396e-06, "loss": 1.0809, "step": 3808 }, { "epoch": 2.551239115874079, "grad_norm": 2.2332007479761913, "learning_rate": 1.7461063294732361e-06, "loss": 0.9644, "step": 3809 }, { "epoch": 2.5519089082384463, "grad_norm": 2.702028591737692, "learning_rate": 1.7447130152951542e-06, "loss": 1.1071, "step": 3810 }, { "epoch": 2.552578700602813, "grad_norm": 3.200977435345308, "learning_rate": 1.743319959233953e-06, "loss": 1.0062, "step": 3811 }, { "epoch": 2.55324849296718, "grad_norm": 2.7132154986336507, "learning_rate": 1.7419271617657065e-06, "loss": 1.1715, "step": 3812 }, { "epoch": 2.553918285331547, "grad_norm": 2.6939145395972828, "learning_rate": 1.740534623366398e-06, "loss": 1.0954, "step": 3813 }, { "epoch": 2.5545880776959144, "grad_norm": 2.8940901984797565, "learning_rate": 1.7391423445119237e-06, "loss": 0.9771, "step": 3814 }, { "epoch": 2.5552578700602813, "grad_norm": 2.677667889566617, "learning_rate": 1.7377503256780886e-06, "loss": 0.9009, "step": 3815 }, { "epoch": 2.555927662424648, "grad_norm": 3.8148091741780052, "learning_rate": 1.7363585673406124e-06, "loss": 1.101, "step": 3816 }, { "epoch": 2.5565974547890153, "grad_norm": 2.752282109719739, "learning_rate": 1.7349670699751225e-06, "loss": 0.9448, "step": 3817 }, { "epoch": 2.5572672471533826, "grad_norm": 2.828876817368765, "learning_rate": 1.7335758340571618e-06, "loss": 1.0335, "step": 3818 }, { "epoch": 2.5579370395177494, "grad_norm": 2.933948035610662, "learning_rate": 1.7321848600621763e-06, "loss": 1.1196, "step": 3819 }, { "epoch": 2.5586068318821167, "grad_norm": 4.119250490100423, "learning_rate": 1.7307941484655297e-06, "loss": 1.0276, "step": 3820 }, { "epoch": 2.5592766242464835, "grad_norm": 5.755712783994555, "learning_rate": 1.7294036997424918e-06, "loss": 1.107, "step": 3821 }, { "epoch": 2.5599464166108508, "grad_norm": 2.5191021776441542, "learning_rate": 1.728013514368246e-06, "loss": 1.0609, "step": 3822 }, { "epoch": 2.5606162089752176, "grad_norm": 2.6354530752096887, "learning_rate": 1.726623592817881e-06, "loss": 1.1466, "step": 3823 }, { "epoch": 2.561286001339585, "grad_norm": 2.487071367975449, "learning_rate": 1.7252339355663997e-06, "loss": 1.0516, "step": 3824 }, { "epoch": 2.5619557937039517, "grad_norm": 2.4809894095125014, "learning_rate": 1.723844543088712e-06, "loss": 1.1504, "step": 3825 }, { "epoch": 2.562625586068319, "grad_norm": 2.5598421453404776, "learning_rate": 1.722455415859639e-06, "loss": 1.1372, "step": 3826 }, { "epoch": 2.5632953784326857, "grad_norm": 2.3516519153231337, "learning_rate": 1.721066554353911e-06, "loss": 0.9575, "step": 3827 }, { "epoch": 2.563965170797053, "grad_norm": 2.5029969852218192, "learning_rate": 1.7196779590461654e-06, "loss": 1.0585, "step": 3828 }, { "epoch": 2.56463496316142, "grad_norm": 2.4411818799792946, "learning_rate": 1.718289630410951e-06, "loss": 1.1348, "step": 3829 }, { "epoch": 2.565304755525787, "grad_norm": 2.6073173947012065, "learning_rate": 1.7169015689227242e-06, "loss": 1.0588, "step": 3830 }, { "epoch": 2.565974547890154, "grad_norm": 2.2965173112164226, "learning_rate": 1.7155137750558517e-06, "loss": 1.0979, "step": 3831 }, { "epoch": 2.566644340254521, "grad_norm": 2.560893403561478, "learning_rate": 1.714126249284606e-06, "loss": 0.937, "step": 3832 }, { "epoch": 2.567314132618888, "grad_norm": 2.604046441541299, "learning_rate": 1.7127389920831699e-06, "loss": 1.1788, "step": 3833 }, { "epoch": 2.5679839249832552, "grad_norm": 2.689076127767074, "learning_rate": 1.7113520039256347e-06, "loss": 1.0935, "step": 3834 }, { "epoch": 2.5686537173476225, "grad_norm": 2.714040465278325, "learning_rate": 1.7099652852859997e-06, "loss": 1.048, "step": 3835 }, { "epoch": 2.5693235097119893, "grad_norm": 4.063996639450928, "learning_rate": 1.7085788366381698e-06, "loss": 1.0429, "step": 3836 }, { "epoch": 2.569993302076356, "grad_norm": 4.808241012376256, "learning_rate": 1.7071926584559595e-06, "loss": 1.0773, "step": 3837 }, { "epoch": 2.5706630944407234, "grad_norm": 2.4188806950074286, "learning_rate": 1.7058067512130922e-06, "loss": 0.9504, "step": 3838 }, { "epoch": 2.5713328868050906, "grad_norm": 2.4136263361060633, "learning_rate": 1.7044211153831966e-06, "loss": 1.0806, "step": 3839 }, { "epoch": 2.5720026791694575, "grad_norm": 3.0935652774674502, "learning_rate": 1.7030357514398085e-06, "loss": 1.1033, "step": 3840 }, { "epoch": 2.5726724715338243, "grad_norm": 2.5492290211942694, "learning_rate": 1.7016506598563715e-06, "loss": 1.0371, "step": 3841 }, { "epoch": 2.5733422638981915, "grad_norm": 3.0323571965967706, "learning_rate": 1.7002658411062373e-06, "loss": 0.9642, "step": 3842 }, { "epoch": 2.574012056262559, "grad_norm": 2.66580889001, "learning_rate": 1.6988812956626633e-06, "loss": 0.9975, "step": 3843 }, { "epoch": 2.5746818486269256, "grad_norm": 3.59423196844631, "learning_rate": 1.697497023998812e-06, "loss": 0.9634, "step": 3844 }, { "epoch": 2.5753516409912924, "grad_norm": 2.3545200908363073, "learning_rate": 1.6961130265877542e-06, "loss": 1.0936, "step": 3845 }, { "epoch": 2.5760214333556597, "grad_norm": 2.1356571064745564, "learning_rate": 1.6947293039024676e-06, "loss": 0.987, "step": 3846 }, { "epoch": 2.576691225720027, "grad_norm": 2.5101307346811517, "learning_rate": 1.6933458564158342e-06, "loss": 1.0583, "step": 3847 }, { "epoch": 2.5773610180843938, "grad_norm": 2.5150663819243335, "learning_rate": 1.691962684600642e-06, "loss": 1.1106, "step": 3848 }, { "epoch": 2.578030810448761, "grad_norm": 2.7961775203560966, "learning_rate": 1.6905797889295866e-06, "loss": 0.8619, "step": 3849 }, { "epoch": 2.578700602813128, "grad_norm": 2.613734083737082, "learning_rate": 1.6891971698752676e-06, "loss": 0.9224, "step": 3850 }, { "epoch": 2.579370395177495, "grad_norm": 3.158514521487666, "learning_rate": 1.6878148279101914e-06, "loss": 0.9521, "step": 3851 }, { "epoch": 2.580040187541862, "grad_norm": 2.853662234795468, "learning_rate": 1.6864327635067673e-06, "loss": 1.0262, "step": 3852 }, { "epoch": 2.580709979906229, "grad_norm": 2.551075522800797, "learning_rate": 1.6850509771373124e-06, "loss": 1.117, "step": 3853 }, { "epoch": 2.581379772270596, "grad_norm": 2.946981382003053, "learning_rate": 1.6836694692740477e-06, "loss": 1.1782, "step": 3854 }, { "epoch": 2.5820495646349633, "grad_norm": 2.743855281899441, "learning_rate": 1.6822882403890995e-06, "loss": 1.0662, "step": 3855 }, { "epoch": 2.58271935699933, "grad_norm": 2.643022844666529, "learning_rate": 1.6809072909544965e-06, "loss": 1.13, "step": 3856 }, { "epoch": 2.5833891493636973, "grad_norm": 3.0980123797809513, "learning_rate": 1.6795266214421752e-06, "loss": 1.0734, "step": 3857 }, { "epoch": 2.584058941728064, "grad_norm": 2.4358897974730365, "learning_rate": 1.678146232323975e-06, "loss": 0.8797, "step": 3858 }, { "epoch": 2.5847287340924314, "grad_norm": 2.876931020040709, "learning_rate": 1.6767661240716381e-06, "loss": 0.9906, "step": 3859 }, { "epoch": 2.5853985264567982, "grad_norm": 2.628709916739044, "learning_rate": 1.6753862971568144e-06, "loss": 1.0467, "step": 3860 }, { "epoch": 2.5860683188211655, "grad_norm": 3.5265688535296884, "learning_rate": 1.6740067520510533e-06, "loss": 0.9335, "step": 3861 }, { "epoch": 2.5867381111855323, "grad_norm": 3.1421579013733085, "learning_rate": 1.6726274892258096e-06, "loss": 0.8573, "step": 3862 }, { "epoch": 2.5874079035498996, "grad_norm": 2.64846389718435, "learning_rate": 1.6712485091524434e-06, "loss": 1.0876, "step": 3863 }, { "epoch": 2.588077695914267, "grad_norm": 2.6456769968249083, "learning_rate": 1.6698698123022168e-06, "loss": 1.2065, "step": 3864 }, { "epoch": 2.5887474882786337, "grad_norm": 3.393236800012308, "learning_rate": 1.6684913991462932e-06, "loss": 0.944, "step": 3865 }, { "epoch": 2.5894172806430005, "grad_norm": 2.774415333150949, "learning_rate": 1.6671132701557414e-06, "loss": 1.0008, "step": 3866 }, { "epoch": 2.5900870730073677, "grad_norm": 2.6807380872102957, "learning_rate": 1.6657354258015334e-06, "loss": 1.0965, "step": 3867 }, { "epoch": 2.590756865371735, "grad_norm": 2.500291898953312, "learning_rate": 1.6643578665545429e-06, "loss": 0.9711, "step": 3868 }, { "epoch": 2.591426657736102, "grad_norm": 2.5643593024168903, "learning_rate": 1.6629805928855457e-06, "loss": 1.0577, "step": 3869 }, { "epoch": 2.5920964501004686, "grad_norm": 3.5096870281856893, "learning_rate": 1.6616036052652208e-06, "loss": 0.9153, "step": 3870 }, { "epoch": 2.592766242464836, "grad_norm": 3.6193906839375036, "learning_rate": 1.6602269041641495e-06, "loss": 1.1159, "step": 3871 }, { "epoch": 2.593436034829203, "grad_norm": 2.7127940976929223, "learning_rate": 1.6588504900528151e-06, "loss": 1.1945, "step": 3872 }, { "epoch": 2.59410582719357, "grad_norm": 3.192498991566579, "learning_rate": 1.6574743634016021e-06, "loss": 1.0491, "step": 3873 }, { "epoch": 2.594775619557937, "grad_norm": 2.4254804832736165, "learning_rate": 1.656098524680797e-06, "loss": 1.0685, "step": 3874 }, { "epoch": 2.595445411922304, "grad_norm": 2.6592097148786515, "learning_rate": 1.6547229743605897e-06, "loss": 1.0685, "step": 3875 }, { "epoch": 2.5961152042866713, "grad_norm": 2.6873780638903786, "learning_rate": 1.6533477129110693e-06, "loss": 0.9702, "step": 3876 }, { "epoch": 2.596784996651038, "grad_norm": 2.6783567571116618, "learning_rate": 1.6519727408022262e-06, "loss": 1.1056, "step": 3877 }, { "epoch": 2.5974547890154054, "grad_norm": 2.3451797161280195, "learning_rate": 1.6505980585039533e-06, "loss": 1.1182, "step": 3878 }, { "epoch": 2.598124581379772, "grad_norm": 2.405122729374788, "learning_rate": 1.649223666486044e-06, "loss": 1.1182, "step": 3879 }, { "epoch": 2.5987943737441395, "grad_norm": 2.555444208834958, "learning_rate": 1.647849565218193e-06, "loss": 1.0144, "step": 3880 }, { "epoch": 2.5994641661085063, "grad_norm": 2.932216151507394, "learning_rate": 1.6464757551699934e-06, "loss": 1.027, "step": 3881 }, { "epoch": 2.6001339584728735, "grad_norm": 2.7764293415822134, "learning_rate": 1.6451022368109404e-06, "loss": 1.045, "step": 3882 }, { "epoch": 2.6008037508372404, "grad_norm": 3.3520913312563145, "learning_rate": 1.6437290106104308e-06, "loss": 1.0171, "step": 3883 }, { "epoch": 2.6014735432016076, "grad_norm": 2.7726750036006402, "learning_rate": 1.64235607703776e-06, "loss": 1.1702, "step": 3884 }, { "epoch": 2.6021433355659744, "grad_norm": 2.811483054847499, "learning_rate": 1.6409834365621225e-06, "loss": 1.0332, "step": 3885 }, { "epoch": 2.6028131279303417, "grad_norm": 3.1019192948241874, "learning_rate": 1.6396110896526141e-06, "loss": 0.9615, "step": 3886 }, { "epoch": 2.6034829202947085, "grad_norm": 3.149276160930594, "learning_rate": 1.6382390367782303e-06, "loss": 1.0395, "step": 3887 }, { "epoch": 2.6041527126590758, "grad_norm": 2.6288504537299024, "learning_rate": 1.6368672784078665e-06, "loss": 1.0931, "step": 3888 }, { "epoch": 2.6048225050234426, "grad_norm": 2.815081617931624, "learning_rate": 1.635495815010315e-06, "loss": 1.2447, "step": 3889 }, { "epoch": 2.60549229738781, "grad_norm": 2.9538071343161363, "learning_rate": 1.6341246470542693e-06, "loss": 1.0189, "step": 3890 }, { "epoch": 2.6061620897521767, "grad_norm": 2.301441316376661, "learning_rate": 1.6327537750083227e-06, "loss": 0.8976, "step": 3891 }, { "epoch": 2.606831882116544, "grad_norm": 2.2367767455461456, "learning_rate": 1.6313831993409662e-06, "loss": 0.8888, "step": 3892 }, { "epoch": 2.607501674480911, "grad_norm": 2.7278625105318803, "learning_rate": 1.6300129205205879e-06, "loss": 1.0334, "step": 3893 }, { "epoch": 2.608171466845278, "grad_norm": 2.799153909666151, "learning_rate": 1.6286429390154778e-06, "loss": 1.0246, "step": 3894 }, { "epoch": 2.608841259209645, "grad_norm": 2.5089764793208937, "learning_rate": 1.627273255293822e-06, "loss": 1.0352, "step": 3895 }, { "epoch": 2.609511051574012, "grad_norm": 2.3914501179328744, "learning_rate": 1.6259038698237054e-06, "loss": 1.0679, "step": 3896 }, { "epoch": 2.6101808439383793, "grad_norm": 2.690278270013716, "learning_rate": 1.6245347830731122e-06, "loss": 1.1646, "step": 3897 }, { "epoch": 2.610850636302746, "grad_norm": 2.5319224643878475, "learning_rate": 1.6231659955099219e-06, "loss": 0.9779, "step": 3898 }, { "epoch": 2.611520428667113, "grad_norm": 3.357957663577677, "learning_rate": 1.6217975076019137e-06, "loss": 0.9569, "step": 3899 }, { "epoch": 2.6121902210314802, "grad_norm": 2.635235593447203, "learning_rate": 1.6204293198167637e-06, "loss": 1.0606, "step": 3900 }, { "epoch": 2.6128600133958475, "grad_norm": 2.7940517032462027, "learning_rate": 1.6190614326220465e-06, "loss": 1.0513, "step": 3901 }, { "epoch": 2.6135298057602143, "grad_norm": 2.955813823178982, "learning_rate": 1.617693846485232e-06, "loss": 1.0649, "step": 3902 }, { "epoch": 2.614199598124581, "grad_norm": 2.8117261415132364, "learning_rate": 1.6163265618736895e-06, "loss": 1.1111, "step": 3903 }, { "epoch": 2.6148693904889484, "grad_norm": 3.096424407377509, "learning_rate": 1.6149595792546825e-06, "loss": 1.2439, "step": 3904 }, { "epoch": 2.6155391828533157, "grad_norm": 2.9236307554433085, "learning_rate": 1.6135928990953747e-06, "loss": 0.9354, "step": 3905 }, { "epoch": 2.6162089752176825, "grad_norm": 2.6515267347821143, "learning_rate": 1.6122265218628235e-06, "loss": 0.9091, "step": 3906 }, { "epoch": 2.6168787675820493, "grad_norm": 3.0005643739534595, "learning_rate": 1.6108604480239837e-06, "loss": 0.9645, "step": 3907 }, { "epoch": 2.6175485599464166, "grad_norm": 3.076955833720499, "learning_rate": 1.6094946780457071e-06, "loss": 0.9292, "step": 3908 }, { "epoch": 2.618218352310784, "grad_norm": 2.5747888465374738, "learning_rate": 1.6081292123947423e-06, "loss": 0.855, "step": 3909 }, { "epoch": 2.6188881446751506, "grad_norm": 2.6306270947770374, "learning_rate": 1.6067640515377308e-06, "loss": 1.0672, "step": 3910 }, { "epoch": 2.619557937039518, "grad_norm": 2.706258199604171, "learning_rate": 1.605399195941212e-06, "loss": 1.0059, "step": 3911 }, { "epoch": 2.6202277294038847, "grad_norm": 2.687937520967902, "learning_rate": 1.6040346460716223e-06, "loss": 1.0581, "step": 3912 }, { "epoch": 2.620897521768252, "grad_norm": 2.473564668616014, "learning_rate": 1.6026704023952922e-06, "loss": 0.8916, "step": 3913 }, { "epoch": 2.621567314132619, "grad_norm": 2.9038971586341913, "learning_rate": 1.601306465378446e-06, "loss": 1.0157, "step": 3914 }, { "epoch": 2.622237106496986, "grad_norm": 2.5923516412446204, "learning_rate": 1.599942835487205e-06, "loss": 1.0902, "step": 3915 }, { "epoch": 2.622906898861353, "grad_norm": 2.4937402700195688, "learning_rate": 1.5985795131875866e-06, "loss": 1.145, "step": 3916 }, { "epoch": 2.62357669122572, "grad_norm": 2.6564462951938888, "learning_rate": 1.5972164989455016e-06, "loss": 0.9821, "step": 3917 }, { "epoch": 2.624246483590087, "grad_norm": 2.4132520957968286, "learning_rate": 1.5958537932267548e-06, "loss": 1.1354, "step": 3918 }, { "epoch": 2.624916275954454, "grad_norm": 3.5648813201690324, "learning_rate": 1.594491396497046e-06, "loss": 1.0252, "step": 3919 }, { "epoch": 2.625586068318821, "grad_norm": 3.041962581093482, "learning_rate": 1.5931293092219715e-06, "loss": 1.0912, "step": 3920 }, { "epoch": 2.6262558606831883, "grad_norm": 3.41737045946107, "learning_rate": 1.59176753186702e-06, "loss": 1.0159, "step": 3921 }, { "epoch": 2.626925653047555, "grad_norm": 2.785701279253533, "learning_rate": 1.5904060648975733e-06, "loss": 1.0487, "step": 3922 }, { "epoch": 2.6275954454119224, "grad_norm": 2.530491428021258, "learning_rate": 1.5890449087789084e-06, "loss": 1.2981, "step": 3923 }, { "epoch": 2.628265237776289, "grad_norm": 2.445897344602925, "learning_rate": 1.5876840639761974e-06, "loss": 1.1011, "step": 3924 }, { "epoch": 2.6289350301406564, "grad_norm": 2.9500832367807126, "learning_rate": 1.5863235309545038e-06, "loss": 1.0909, "step": 3925 }, { "epoch": 2.6296048225050237, "grad_norm": 2.5368220754110427, "learning_rate": 1.5849633101787853e-06, "loss": 1.0329, "step": 3926 }, { "epoch": 2.6302746148693905, "grad_norm": 2.5293583937333177, "learning_rate": 1.583603402113892e-06, "loss": 1.1866, "step": 3927 }, { "epoch": 2.6309444072337573, "grad_norm": 3.2118317731645396, "learning_rate": 1.5822438072245699e-06, "loss": 0.9248, "step": 3928 }, { "epoch": 2.6316141995981246, "grad_norm": 2.4450670494156093, "learning_rate": 1.5808845259754557e-06, "loss": 0.9109, "step": 3929 }, { "epoch": 2.632283991962492, "grad_norm": 2.334931934573242, "learning_rate": 1.5795255588310787e-06, "loss": 0.9472, "step": 3930 }, { "epoch": 2.6329537843268587, "grad_norm": 2.7496333538718667, "learning_rate": 1.5781669062558607e-06, "loss": 1.009, "step": 3931 }, { "epoch": 2.6336235766912255, "grad_norm": 3.647088691158087, "learning_rate": 1.576808568714119e-06, "loss": 1.0915, "step": 3932 }, { "epoch": 2.6342933690555927, "grad_norm": 2.458235034962863, "learning_rate": 1.5754505466700593e-06, "loss": 1.1872, "step": 3933 }, { "epoch": 2.63496316141996, "grad_norm": 3.005336558753139, "learning_rate": 1.5740928405877828e-06, "loss": 1.0898, "step": 3934 }, { "epoch": 2.635632953784327, "grad_norm": 2.624667383693507, "learning_rate": 1.5727354509312797e-06, "loss": 1.1186, "step": 3935 }, { "epoch": 2.6363027461486936, "grad_norm": 2.685024780147098, "learning_rate": 1.5713783781644343e-06, "loss": 1.0078, "step": 3936 }, { "epoch": 2.636972538513061, "grad_norm": 2.6935951842851775, "learning_rate": 1.5700216227510212e-06, "loss": 1.0693, "step": 3937 }, { "epoch": 2.637642330877428, "grad_norm": 3.374659348650494, "learning_rate": 1.5686651851547084e-06, "loss": 1.1052, "step": 3938 }, { "epoch": 2.638312123241795, "grad_norm": 3.0715422419841207, "learning_rate": 1.5673090658390528e-06, "loss": 1.2491, "step": 3939 }, { "epoch": 2.6389819156061622, "grad_norm": 2.6189878849166974, "learning_rate": 1.5659532652675041e-06, "loss": 0.9232, "step": 3940 }, { "epoch": 2.639651707970529, "grad_norm": 2.9099921199428462, "learning_rate": 1.564597783903402e-06, "loss": 1.0415, "step": 3941 }, { "epoch": 2.6403215003348963, "grad_norm": 3.802876351009102, "learning_rate": 1.5632426222099807e-06, "loss": 0.9683, "step": 3942 }, { "epoch": 2.640991292699263, "grad_norm": 2.600512953122064, "learning_rate": 1.561887780650359e-06, "loss": 1.053, "step": 3943 }, { "epoch": 2.6416610850636304, "grad_norm": 2.6092593028658664, "learning_rate": 1.5605332596875514e-06, "loss": 1.1791, "step": 3944 }, { "epoch": 2.642330877427997, "grad_norm": 2.597219222228476, "learning_rate": 1.55917905978446e-06, "loss": 1.044, "step": 3945 }, { "epoch": 2.6430006697923645, "grad_norm": 2.872748458972119, "learning_rate": 1.55782518140388e-06, "loss": 1.0089, "step": 3946 }, { "epoch": 2.6436704621567313, "grad_norm": 2.5442359576201663, "learning_rate": 1.5564716250084933e-06, "loss": 0.9041, "step": 3947 }, { "epoch": 2.6443402545210986, "grad_norm": 2.466516898613927, "learning_rate": 1.5551183910608738e-06, "loss": 1.0845, "step": 3948 }, { "epoch": 2.6450100468854654, "grad_norm": 2.921372326947498, "learning_rate": 1.5537654800234846e-06, "loss": 0.9343, "step": 3949 }, { "epoch": 2.6456798392498326, "grad_norm": 3.367188830314396, "learning_rate": 1.55241289235868e-06, "loss": 1.0277, "step": 3950 }, { "epoch": 2.6463496316141994, "grad_norm": 3.386977252928069, "learning_rate": 1.5510606285287008e-06, "loss": 1.0917, "step": 3951 }, { "epoch": 2.6470194239785667, "grad_norm": 2.8856827863292995, "learning_rate": 1.5497086889956791e-06, "loss": 0.9938, "step": 3952 }, { "epoch": 2.6476892163429335, "grad_norm": 3.0699548463575272, "learning_rate": 1.548357074221637e-06, "loss": 0.9529, "step": 3953 }, { "epoch": 2.648359008707301, "grad_norm": 2.830106530561138, "learning_rate": 1.547005784668484e-06, "loss": 0.8943, "step": 3954 }, { "epoch": 2.649028801071668, "grad_norm": 2.6592636948520534, "learning_rate": 1.5456548207980176e-06, "loss": 1.0972, "step": 3955 }, { "epoch": 2.649698593436035, "grad_norm": 5.399704987810429, "learning_rate": 1.5443041830719262e-06, "loss": 0.7885, "step": 3956 }, { "epoch": 2.6503683858004017, "grad_norm": 3.6599931622130213, "learning_rate": 1.5429538719517867e-06, "loss": 0.9363, "step": 3957 }, { "epoch": 2.651038178164769, "grad_norm": 3.219804261155192, "learning_rate": 1.5416038878990636e-06, "loss": 0.9402, "step": 3958 }, { "epoch": 2.651707970529136, "grad_norm": 3.555199950213905, "learning_rate": 1.5402542313751079e-06, "loss": 1.0683, "step": 3959 }, { "epoch": 2.652377762893503, "grad_norm": 2.495025151093635, "learning_rate": 1.5389049028411613e-06, "loss": 1.0161, "step": 3960 }, { "epoch": 2.65304755525787, "grad_norm": 3.278250854942268, "learning_rate": 1.5375559027583524e-06, "loss": 0.9849, "step": 3961 }, { "epoch": 2.653717347622237, "grad_norm": 2.6376086593390506, "learning_rate": 1.5362072315876986e-06, "loss": 1.05, "step": 3962 }, { "epoch": 2.6543871399866044, "grad_norm": 2.703961541710898, "learning_rate": 1.5348588897901021e-06, "loss": 1.0759, "step": 3963 }, { "epoch": 2.655056932350971, "grad_norm": 2.7272897029826986, "learning_rate": 1.533510877826355e-06, "loss": 1.0853, "step": 3964 }, { "epoch": 2.655726724715338, "grad_norm": 2.490911638383466, "learning_rate": 1.5321631961571365e-06, "loss": 0.9314, "step": 3965 }, { "epoch": 2.6563965170797053, "grad_norm": 2.69164575625498, "learning_rate": 1.5308158452430119e-06, "loss": 1.092, "step": 3966 }, { "epoch": 2.6570663094440725, "grad_norm": 2.686910375086917, "learning_rate": 1.5294688255444357e-06, "loss": 0.9735, "step": 3967 }, { "epoch": 2.6577361018084393, "grad_norm": 2.593752707644624, "learning_rate": 1.5281221375217447e-06, "loss": 0.9743, "step": 3968 }, { "epoch": 2.6584058941728066, "grad_norm": 2.6083057258187012, "learning_rate": 1.5267757816351672e-06, "loss": 0.9151, "step": 3969 }, { "epoch": 2.6590756865371734, "grad_norm": 2.703864730900411, "learning_rate": 1.5254297583448146e-06, "loss": 1.0453, "step": 3970 }, { "epoch": 2.6597454789015407, "grad_norm": 4.231453696231113, "learning_rate": 1.5240840681106885e-06, "loss": 0.9064, "step": 3971 }, { "epoch": 2.6604152712659075, "grad_norm": 3.2730859150539375, "learning_rate": 1.5227387113926709e-06, "loss": 0.8695, "step": 3972 }, { "epoch": 2.6610850636302747, "grad_norm": 2.7271341100008653, "learning_rate": 1.5213936886505352e-06, "loss": 1.1297, "step": 3973 }, { "epoch": 2.6617548559946416, "grad_norm": 2.37398908409686, "learning_rate": 1.5200490003439372e-06, "loss": 0.9014, "step": 3974 }, { "epoch": 2.662424648359009, "grad_norm": 2.7533730925075077, "learning_rate": 1.5187046469324224e-06, "loss": 1.1037, "step": 3975 }, { "epoch": 2.6630944407233756, "grad_norm": 2.505925936458009, "learning_rate": 1.5173606288754162e-06, "loss": 1.1549, "step": 3976 }, { "epoch": 2.663764233087743, "grad_norm": 2.8518652322930333, "learning_rate": 1.5160169466322333e-06, "loss": 1.0432, "step": 3977 }, { "epoch": 2.6644340254521097, "grad_norm": 4.510018793425778, "learning_rate": 1.5146736006620732e-06, "loss": 1.1844, "step": 3978 }, { "epoch": 2.665103817816477, "grad_norm": 2.7682709804487415, "learning_rate": 1.513330591424021e-06, "loss": 1.0846, "step": 3979 }, { "epoch": 2.665773610180844, "grad_norm": 2.4930076385887294, "learning_rate": 1.5119879193770438e-06, "loss": 1.0447, "step": 3980 }, { "epoch": 2.666443402545211, "grad_norm": 2.9317449296476044, "learning_rate": 1.5106455849799968e-06, "loss": 0.9404, "step": 3981 }, { "epoch": 2.667113194909578, "grad_norm": 3.4770469315942214, "learning_rate": 1.5093035886916168e-06, "loss": 0.7755, "step": 3982 }, { "epoch": 2.667782987273945, "grad_norm": 2.6455596516547266, "learning_rate": 1.5079619309705293e-06, "loss": 1.0563, "step": 3983 }, { "epoch": 2.6684527796383124, "grad_norm": 2.501702413599401, "learning_rate": 1.5066206122752395e-06, "loss": 1.1451, "step": 3984 }, { "epoch": 2.669122572002679, "grad_norm": 6.170603066139006, "learning_rate": 1.505279633064139e-06, "loss": 1.0464, "step": 3985 }, { "epoch": 2.669792364367046, "grad_norm": 2.4201544204437715, "learning_rate": 1.5039389937955029e-06, "loss": 0.9388, "step": 3986 }, { "epoch": 2.6704621567314133, "grad_norm": 3.0959403763320115, "learning_rate": 1.5025986949274921e-06, "loss": 1.0004, "step": 3987 }, { "epoch": 2.6711319490957806, "grad_norm": 6.522069641246981, "learning_rate": 1.5012587369181472e-06, "loss": 0.926, "step": 3988 }, { "epoch": 2.6718017414601474, "grad_norm": 2.8503013492968616, "learning_rate": 1.4999191202253954e-06, "loss": 0.8146, "step": 3989 }, { "epoch": 2.672471533824514, "grad_norm": 3.068565543647216, "learning_rate": 1.498579845307046e-06, "loss": 1.1912, "step": 3990 }, { "epoch": 2.6731413261888815, "grad_norm": 2.8383583621792243, "learning_rate": 1.4972409126207932e-06, "loss": 1.0853, "step": 3991 }, { "epoch": 2.6738111185532487, "grad_norm": 2.909175404679589, "learning_rate": 1.4959023226242115e-06, "loss": 1.1459, "step": 3992 }, { "epoch": 2.6744809109176155, "grad_norm": 2.835158626731869, "learning_rate": 1.4945640757747601e-06, "loss": 1.1315, "step": 3993 }, { "epoch": 2.6751507032819823, "grad_norm": 3.0026362640806363, "learning_rate": 1.4932261725297797e-06, "loss": 1.116, "step": 3994 }, { "epoch": 2.6758204956463496, "grad_norm": 3.461467603312619, "learning_rate": 1.4918886133464966e-06, "loss": 1.0311, "step": 3995 }, { "epoch": 2.676490288010717, "grad_norm": 2.8887434145485593, "learning_rate": 1.4905513986820155e-06, "loss": 1.0526, "step": 3996 }, { "epoch": 2.6771600803750837, "grad_norm": 2.654475713010402, "learning_rate": 1.4892145289933251e-06, "loss": 1.0539, "step": 3997 }, { "epoch": 2.6778298727394505, "grad_norm": 2.798182667488288, "learning_rate": 1.4878780047372973e-06, "loss": 1.1191, "step": 3998 }, { "epoch": 2.6784996651038178, "grad_norm": 2.606930099454349, "learning_rate": 1.4865418263706849e-06, "loss": 1.1418, "step": 3999 }, { "epoch": 2.679169457468185, "grad_norm": 2.4704959606279195, "learning_rate": 1.4852059943501218e-06, "loss": 0.9921, "step": 4000 }, { "epoch": 2.679839249832552, "grad_norm": 3.442239472375468, "learning_rate": 1.483870509132124e-06, "loss": 0.9557, "step": 4001 }, { "epoch": 2.680509042196919, "grad_norm": 2.5288477012175936, "learning_rate": 1.4825353711730906e-06, "loss": 1.0579, "step": 4002 }, { "epoch": 2.681178834561286, "grad_norm": 2.9627217547590066, "learning_rate": 1.4812005809292994e-06, "loss": 1.0381, "step": 4003 }, { "epoch": 2.681848626925653, "grad_norm": 4.261081277161753, "learning_rate": 1.4798661388569119e-06, "loss": 0.8964, "step": 4004 }, { "epoch": 2.68251841929002, "grad_norm": 2.5031891414853877, "learning_rate": 1.4785320454119676e-06, "loss": 1.0823, "step": 4005 }, { "epoch": 2.6831882116543873, "grad_norm": 2.5969513487625497, "learning_rate": 1.4771983010503899e-06, "loss": 1.0064, "step": 4006 }, { "epoch": 2.683858004018754, "grad_norm": 3.0014791420875158, "learning_rate": 1.4758649062279818e-06, "loss": 1.0451, "step": 4007 }, { "epoch": 2.6845277963831213, "grad_norm": 6.311647964285789, "learning_rate": 1.4745318614004264e-06, "loss": 0.7976, "step": 4008 }, { "epoch": 2.685197588747488, "grad_norm": 3.4991994901444095, "learning_rate": 1.4731991670232862e-06, "loss": 1.1117, "step": 4009 }, { "epoch": 2.6858673811118554, "grad_norm": 2.755336634132871, "learning_rate": 1.4718668235520068e-06, "loss": 1.1123, "step": 4010 }, { "epoch": 2.6865371734762222, "grad_norm": 2.4932583182931647, "learning_rate": 1.4705348314419115e-06, "loss": 0.9367, "step": 4011 }, { "epoch": 2.6872069658405895, "grad_norm": 3.0660618268810493, "learning_rate": 1.4692031911482057e-06, "loss": 1.1068, "step": 4012 }, { "epoch": 2.6878767582049563, "grad_norm": 2.519382086874184, "learning_rate": 1.4678719031259703e-06, "loss": 1.074, "step": 4013 }, { "epoch": 2.6885465505693236, "grad_norm": 3.257303037069432, "learning_rate": 1.4665409678301712e-06, "loss": 1.0752, "step": 4014 }, { "epoch": 2.6892163429336904, "grad_norm": 2.5935546451468263, "learning_rate": 1.4652103857156497e-06, "loss": 1.0384, "step": 4015 }, { "epoch": 2.6898861352980576, "grad_norm": 2.7826348194269905, "learning_rate": 1.4638801572371303e-06, "loss": 0.9636, "step": 4016 }, { "epoch": 2.690555927662425, "grad_norm": 2.56164476937795, "learning_rate": 1.4625502828492107e-06, "loss": 1.0758, "step": 4017 }, { "epoch": 2.6912257200267917, "grad_norm": 2.799662592319332, "learning_rate": 1.4612207630063738e-06, "loss": 1.0816, "step": 4018 }, { "epoch": 2.6918955123911585, "grad_norm": 2.685644725832736, "learning_rate": 1.4598915981629774e-06, "loss": 0.9565, "step": 4019 }, { "epoch": 2.692565304755526, "grad_norm": 3.057626492639091, "learning_rate": 1.4585627887732606e-06, "loss": 0.734, "step": 4020 }, { "epoch": 2.693235097119893, "grad_norm": 2.58428960220757, "learning_rate": 1.4572343352913388e-06, "loss": 1.006, "step": 4021 }, { "epoch": 2.69390488948426, "grad_norm": 2.821002130637527, "learning_rate": 1.4559062381712066e-06, "loss": 0.9528, "step": 4022 }, { "epoch": 2.6945746818486267, "grad_norm": 3.337079566528955, "learning_rate": 1.454578497866737e-06, "loss": 1.0115, "step": 4023 }, { "epoch": 2.695244474212994, "grad_norm": 2.4853348542630274, "learning_rate": 1.4532511148316814e-06, "loss": 1.0015, "step": 4024 }, { "epoch": 2.695914266577361, "grad_norm": 2.9022897110279513, "learning_rate": 1.451924089519668e-06, "loss": 1.0789, "step": 4025 }, { "epoch": 2.696584058941728, "grad_norm": 2.9336329585136505, "learning_rate": 1.4505974223842037e-06, "loss": 1.0043, "step": 4026 }, { "epoch": 2.697253851306095, "grad_norm": 2.402246812593803, "learning_rate": 1.4492711138786728e-06, "loss": 0.9839, "step": 4027 }, { "epoch": 2.697923643670462, "grad_norm": 2.5088357769464413, "learning_rate": 1.4479451644563367e-06, "loss": 1.0109, "step": 4028 }, { "epoch": 2.6985934360348294, "grad_norm": 2.7069436461531864, "learning_rate": 1.4466195745703345e-06, "loss": 0.8232, "step": 4029 }, { "epoch": 2.699263228399196, "grad_norm": 3.197338306830084, "learning_rate": 1.4452943446736823e-06, "loss": 1.0672, "step": 4030 }, { "epoch": 2.6999330207635635, "grad_norm": 2.800862717880337, "learning_rate": 1.4439694752192729e-06, "loss": 0.9621, "step": 4031 }, { "epoch": 2.7006028131279303, "grad_norm": 2.9479394446253666, "learning_rate": 1.4426449666598764e-06, "loss": 1.0475, "step": 4032 }, { "epoch": 2.7012726054922975, "grad_norm": 3.280717675032003, "learning_rate": 1.4413208194481392e-06, "loss": 1.0588, "step": 4033 }, { "epoch": 2.7019423978566643, "grad_norm": 3.132010007982803, "learning_rate": 1.4399970340365842e-06, "loss": 1.1809, "step": 4034 }, { "epoch": 2.7026121902210316, "grad_norm": 2.855048917362611, "learning_rate": 1.4386736108776111e-06, "loss": 0.9839, "step": 4035 }, { "epoch": 2.7032819825853984, "grad_norm": 2.7048342512540513, "learning_rate": 1.4373505504234953e-06, "loss": 1.0357, "step": 4036 }, { "epoch": 2.7039517749497657, "grad_norm": 3.834469200802941, "learning_rate": 1.4360278531263883e-06, "loss": 1.0477, "step": 4037 }, { "epoch": 2.7046215673141325, "grad_norm": 2.7916635884112018, "learning_rate": 1.4347055194383175e-06, "loss": 1.0867, "step": 4038 }, { "epoch": 2.7052913596784998, "grad_norm": 3.7238213680424015, "learning_rate": 1.4333835498111868e-06, "loss": 1.0713, "step": 4039 }, { "epoch": 2.7059611520428666, "grad_norm": 2.9781461988705424, "learning_rate": 1.4320619446967732e-06, "loss": 1.2552, "step": 4040 }, { "epoch": 2.706630944407234, "grad_norm": 2.7449344797552895, "learning_rate": 1.4307407045467348e-06, "loss": 0.9129, "step": 4041 }, { "epoch": 2.7073007367716007, "grad_norm": 2.7121585697661543, "learning_rate": 1.429419829812597e-06, "loss": 1.1382, "step": 4042 }, { "epoch": 2.707970529135968, "grad_norm": 2.5711291492142205, "learning_rate": 1.4280993209457671e-06, "loss": 0.8965, "step": 4043 }, { "epoch": 2.7086403215003347, "grad_norm": 2.5501047008340234, "learning_rate": 1.4267791783975224e-06, "loss": 0.9795, "step": 4044 }, { "epoch": 2.709310113864702, "grad_norm": 3.1132289207587607, "learning_rate": 1.4254594026190208e-06, "loss": 0.924, "step": 4045 }, { "epoch": 2.7099799062290693, "grad_norm": 2.274637680891531, "learning_rate": 1.4241399940612882e-06, "loss": 0.9477, "step": 4046 }, { "epoch": 2.710649698593436, "grad_norm": 2.805448292080407, "learning_rate": 1.4228209531752285e-06, "loss": 1.0415, "step": 4047 }, { "epoch": 2.711319490957803, "grad_norm": 2.55013495247662, "learning_rate": 1.4215022804116219e-06, "loss": 1.0607, "step": 4048 }, { "epoch": 2.71198928332217, "grad_norm": 2.886895047891588, "learning_rate": 1.4201839762211197e-06, "loss": 0.7878, "step": 4049 }, { "epoch": 2.7126590756865374, "grad_norm": 3.64960965408196, "learning_rate": 1.4188660410542466e-06, "loss": 1.0657, "step": 4050 }, { "epoch": 2.7133288680509042, "grad_norm": 2.7000725890395905, "learning_rate": 1.4175484753614027e-06, "loss": 0.8858, "step": 4051 }, { "epoch": 2.713998660415271, "grad_norm": 2.706584050294304, "learning_rate": 1.416231279592864e-06, "loss": 1.0331, "step": 4052 }, { "epoch": 2.7146684527796383, "grad_norm": 2.810745604964085, "learning_rate": 1.4149144541987772e-06, "loss": 1.0124, "step": 4053 }, { "epoch": 2.7153382451440056, "grad_norm": 3.75726680212827, "learning_rate": 1.4135979996291618e-06, "loss": 1.2386, "step": 4054 }, { "epoch": 2.7160080375083724, "grad_norm": 3.1452204512827593, "learning_rate": 1.4122819163339117e-06, "loss": 1.1626, "step": 4055 }, { "epoch": 2.716677829872739, "grad_norm": 4.098908155248428, "learning_rate": 1.4109662047627959e-06, "loss": 0.897, "step": 4056 }, { "epoch": 2.7173476222371065, "grad_norm": 3.1743849061433886, "learning_rate": 1.4096508653654545e-06, "loss": 0.9898, "step": 4057 }, { "epoch": 2.7180174146014737, "grad_norm": 3.38821952623057, "learning_rate": 1.4083358985913992e-06, "loss": 1.0264, "step": 4058 }, { "epoch": 2.7186872069658405, "grad_norm": 2.4749598577096186, "learning_rate": 1.4070213048900146e-06, "loss": 1.1238, "step": 4059 }, { "epoch": 2.7193569993302074, "grad_norm": 5.557149464800545, "learning_rate": 1.405707084710562e-06, "loss": 1.1619, "step": 4060 }, { "epoch": 2.7200267916945746, "grad_norm": 2.917469140325391, "learning_rate": 1.4043932385021708e-06, "loss": 1.002, "step": 4061 }, { "epoch": 2.720696584058942, "grad_norm": 3.1038186940929244, "learning_rate": 1.4030797667138421e-06, "loss": 1.118, "step": 4062 }, { "epoch": 2.7213663764233087, "grad_norm": 3.1043904558148423, "learning_rate": 1.4017666697944506e-06, "loss": 1.081, "step": 4063 }, { "epoch": 2.722036168787676, "grad_norm": 2.69073482847381, "learning_rate": 1.4004539481927453e-06, "loss": 0.9864, "step": 4064 }, { "epoch": 2.7227059611520428, "grad_norm": 3.018244095918385, "learning_rate": 1.3991416023573442e-06, "loss": 0.9418, "step": 4065 }, { "epoch": 2.72337575351641, "grad_norm": 2.7065060409336006, "learning_rate": 1.3978296327367352e-06, "loss": 1.2126, "step": 4066 }, { "epoch": 2.724045545880777, "grad_norm": 2.772052570997929, "learning_rate": 1.3965180397792798e-06, "loss": 1.0207, "step": 4067 }, { "epoch": 2.724715338245144, "grad_norm": 2.6901278358178007, "learning_rate": 1.395206823933213e-06, "loss": 1.1406, "step": 4068 }, { "epoch": 2.725385130609511, "grad_norm": 3.0189318710166946, "learning_rate": 1.3938959856466377e-06, "loss": 1.1909, "step": 4069 }, { "epoch": 2.726054922973878, "grad_norm": 4.502101622044999, "learning_rate": 1.392585525367527e-06, "loss": 0.9408, "step": 4070 }, { "epoch": 2.726724715338245, "grad_norm": 3.1104697978692863, "learning_rate": 1.391275443543727e-06, "loss": 0.9615, "step": 4071 }, { "epoch": 2.7273945077026123, "grad_norm": 2.9045244051500845, "learning_rate": 1.3899657406229555e-06, "loss": 0.8261, "step": 4072 }, { "epoch": 2.728064300066979, "grad_norm": 2.722512586780823, "learning_rate": 1.3886564170527989e-06, "loss": 1.0229, "step": 4073 }, { "epoch": 2.7287340924313463, "grad_norm": 2.84703625186651, "learning_rate": 1.387347473280713e-06, "loss": 0.8252, "step": 4074 }, { "epoch": 2.729403884795713, "grad_norm": 2.773502610812234, "learning_rate": 1.3860389097540244e-06, "loss": 1.0943, "step": 4075 }, { "epoch": 2.7300736771600804, "grad_norm": 2.8411766131221134, "learning_rate": 1.3847307269199328e-06, "loss": 0.8057, "step": 4076 }, { "epoch": 2.7307434695244472, "grad_norm": 2.578793687338223, "learning_rate": 1.3834229252255047e-06, "loss": 1.1038, "step": 4077 }, { "epoch": 2.7314132618888145, "grad_norm": 2.6140460557487804, "learning_rate": 1.3821155051176765e-06, "loss": 0.9814, "step": 4078 }, { "epoch": 2.7320830542531818, "grad_norm": 2.984205655590565, "learning_rate": 1.3808084670432553e-06, "loss": 1.102, "step": 4079 }, { "epoch": 2.7327528466175486, "grad_norm": 2.834404339887753, "learning_rate": 1.379501811448917e-06, "loss": 1.0184, "step": 4080 }, { "epoch": 2.7334226389819154, "grad_norm": 2.8463565124707286, "learning_rate": 1.3781955387812067e-06, "loss": 1.0723, "step": 4081 }, { "epoch": 2.7340924313462827, "grad_norm": 3.0028338334843, "learning_rate": 1.376889649486539e-06, "loss": 1.2204, "step": 4082 }, { "epoch": 2.73476222371065, "grad_norm": 2.599493581136908, "learning_rate": 1.3755841440111974e-06, "loss": 1.0085, "step": 4083 }, { "epoch": 2.7354320160750167, "grad_norm": 2.6892624949041104, "learning_rate": 1.374279022801334e-06, "loss": 1.1163, "step": 4084 }, { "epoch": 2.7361018084393836, "grad_norm": 2.527269287461678, "learning_rate": 1.3729742863029703e-06, "loss": 1.1666, "step": 4085 }, { "epoch": 2.736771600803751, "grad_norm": 2.794062292815226, "learning_rate": 1.371669934961995e-06, "loss": 0.8564, "step": 4086 }, { "epoch": 2.737441393168118, "grad_norm": 3.645479136549046, "learning_rate": 1.3703659692241662e-06, "loss": 0.9723, "step": 4087 }, { "epoch": 2.738111185532485, "grad_norm": 4.307375957059567, "learning_rate": 1.36906238953511e-06, "loss": 0.8395, "step": 4088 }, { "epoch": 2.7387809778968517, "grad_norm": 3.0990006546960474, "learning_rate": 1.3677591963403207e-06, "loss": 1.0893, "step": 4089 }, { "epoch": 2.739450770261219, "grad_norm": 2.927451351279929, "learning_rate": 1.3664563900851604e-06, "loss": 1.1362, "step": 4090 }, { "epoch": 2.7401205626255862, "grad_norm": 2.9064238570654903, "learning_rate": 1.3651539712148587e-06, "loss": 1.0132, "step": 4091 }, { "epoch": 2.740790354989953, "grad_norm": 2.233843807491158, "learning_rate": 1.3638519401745133e-06, "loss": 0.7842, "step": 4092 }, { "epoch": 2.7414601473543203, "grad_norm": 2.8917279861408267, "learning_rate": 1.3625502974090884e-06, "loss": 0.9434, "step": 4093 }, { "epoch": 2.742129939718687, "grad_norm": 2.9304645585506828, "learning_rate": 1.361249043363417e-06, "loss": 0.9955, "step": 4094 }, { "epoch": 2.7427997320830544, "grad_norm": 2.635424062151423, "learning_rate": 1.3599481784821982e-06, "loss": 0.9864, "step": 4095 }, { "epoch": 2.743469524447421, "grad_norm": 2.6086912923635315, "learning_rate": 1.3586477032099982e-06, "loss": 1.1216, "step": 4096 }, { "epoch": 2.7441393168117885, "grad_norm": 5.2941946224094005, "learning_rate": 1.3573476179912504e-06, "loss": 0.9806, "step": 4097 }, { "epoch": 2.7448091091761553, "grad_norm": 2.347993207522336, "learning_rate": 1.3560479232702544e-06, "loss": 1.1861, "step": 4098 }, { "epoch": 2.7454789015405225, "grad_norm": 2.614488605546461, "learning_rate": 1.3547486194911775e-06, "loss": 1.035, "step": 4099 }, { "epoch": 2.7461486939048894, "grad_norm": 3.5558936091850457, "learning_rate": 1.3534497070980513e-06, "loss": 0.9503, "step": 4100 }, { "epoch": 2.7468184862692566, "grad_norm": 3.95529046651248, "learning_rate": 1.3521511865347758e-06, "loss": 1.0372, "step": 4101 }, { "epoch": 2.7474882786336234, "grad_norm": 2.565847478300028, "learning_rate": 1.3508530582451163e-06, "loss": 1.0439, "step": 4102 }, { "epoch": 2.7481580709979907, "grad_norm": 2.688598591865103, "learning_rate": 1.349555322672703e-06, "loss": 1.1462, "step": 4103 }, { "epoch": 2.7488278633623575, "grad_norm": 2.647303600266698, "learning_rate": 1.3482579802610339e-06, "loss": 1.0127, "step": 4104 }, { "epoch": 2.7494976557267248, "grad_norm": 3.18132892597526, "learning_rate": 1.346961031453471e-06, "loss": 1.0051, "step": 4105 }, { "epoch": 2.7501674480910916, "grad_norm": 3.235863677620923, "learning_rate": 1.3456644766932422e-06, "loss": 1.0183, "step": 4106 }, { "epoch": 2.750837240455459, "grad_norm": 2.7398997008315082, "learning_rate": 1.3443683164234413e-06, "loss": 1.0436, "step": 4107 }, { "epoch": 2.751507032819826, "grad_norm": 2.513309619197611, "learning_rate": 1.3430725510870269e-06, "loss": 1.0223, "step": 4108 }, { "epoch": 2.752176825184193, "grad_norm": 2.5112499814952094, "learning_rate": 1.3417771811268224e-06, "loss": 0.985, "step": 4109 }, { "epoch": 2.7528466175485597, "grad_norm": 2.941264661932618, "learning_rate": 1.340482206985515e-06, "loss": 1.2765, "step": 4110 }, { "epoch": 2.753516409912927, "grad_norm": 2.8695740087696904, "learning_rate": 1.3391876291056616e-06, "loss": 0.9366, "step": 4111 }, { "epoch": 2.7541862022772943, "grad_norm": 3.230132214194702, "learning_rate": 1.3378934479296769e-06, "loss": 0.9362, "step": 4112 }, { "epoch": 2.754855994641661, "grad_norm": 2.294172000420249, "learning_rate": 1.3365996638998433e-06, "loss": 0.9293, "step": 4113 }, { "epoch": 2.755525787006028, "grad_norm": 2.52743758539092, "learning_rate": 1.335306277458307e-06, "loss": 0.8659, "step": 4114 }, { "epoch": 2.756195579370395, "grad_norm": 2.4623840436052373, "learning_rate": 1.3340132890470816e-06, "loss": 0.7489, "step": 4115 }, { "epoch": 2.7568653717347624, "grad_norm": 3.46352734418809, "learning_rate": 1.3327206991080388e-06, "loss": 0.905, "step": 4116 }, { "epoch": 2.7575351640991292, "grad_norm": 2.81092883765921, "learning_rate": 1.3314285080829175e-06, "loss": 0.9377, "step": 4117 }, { "epoch": 2.758204956463496, "grad_norm": 3.0978222172324386, "learning_rate": 1.3301367164133194e-06, "loss": 1.0042, "step": 4118 }, { "epoch": 2.7588747488278633, "grad_norm": 2.560361546076466, "learning_rate": 1.3288453245407126e-06, "loss": 0.9789, "step": 4119 }, { "epoch": 2.7595445411922306, "grad_norm": 2.652876278660899, "learning_rate": 1.3275543329064237e-06, "loss": 1.0295, "step": 4120 }, { "epoch": 2.7602143335565974, "grad_norm": 2.648090458016608, "learning_rate": 1.3262637419516453e-06, "loss": 0.9738, "step": 4121 }, { "epoch": 2.7608841259209647, "grad_norm": 2.771953236800215, "learning_rate": 1.3249735521174318e-06, "loss": 1.0628, "step": 4122 }, { "epoch": 2.7615539182853315, "grad_norm": 2.456612300252037, "learning_rate": 1.3236837638447047e-06, "loss": 1.0817, "step": 4123 }, { "epoch": 2.7622237106496987, "grad_norm": 3.5700226682449046, "learning_rate": 1.322394377574242e-06, "loss": 0.9972, "step": 4124 }, { "epoch": 2.7628935030140656, "grad_norm": 2.441296386655441, "learning_rate": 1.3211053937466878e-06, "loss": 0.9998, "step": 4125 }, { "epoch": 2.763563295378433, "grad_norm": 2.8941948611959654, "learning_rate": 1.3198168128025477e-06, "loss": 1.1388, "step": 4126 }, { "epoch": 2.7642330877427996, "grad_norm": 2.94610043795842, "learning_rate": 1.3185286351821925e-06, "loss": 1.0511, "step": 4127 }, { "epoch": 2.764902880107167, "grad_norm": 2.5291198869804967, "learning_rate": 1.3172408613258504e-06, "loss": 0.9144, "step": 4128 }, { "epoch": 2.7655726724715337, "grad_norm": 2.530986780105137, "learning_rate": 1.3159534916736143e-06, "loss": 0.8929, "step": 4129 }, { "epoch": 2.766242464835901, "grad_norm": 3.472214903618284, "learning_rate": 1.314666526665438e-06, "loss": 1.0817, "step": 4130 }, { "epoch": 2.766912257200268, "grad_norm": 2.77702914003802, "learning_rate": 1.3133799667411404e-06, "loss": 1.1172, "step": 4131 }, { "epoch": 2.767582049564635, "grad_norm": 2.686036726705977, "learning_rate": 1.3120938123403964e-06, "loss": 1.0728, "step": 4132 }, { "epoch": 2.768251841929002, "grad_norm": 2.6809972430369484, "learning_rate": 1.310808063902746e-06, "loss": 0.9121, "step": 4133 }, { "epoch": 2.768921634293369, "grad_norm": 3.2487042900516006, "learning_rate": 1.3095227218675886e-06, "loss": 0.9921, "step": 4134 }, { "epoch": 2.769591426657736, "grad_norm": 2.8477518279897844, "learning_rate": 1.3082377866741886e-06, "loss": 1.1509, "step": 4135 }, { "epoch": 2.770261219022103, "grad_norm": 2.7298563046085667, "learning_rate": 1.3069532587616654e-06, "loss": 1.0388, "step": 4136 }, { "epoch": 2.7709310113864705, "grad_norm": 3.404530431855257, "learning_rate": 1.3056691385690023e-06, "loss": 1.0213, "step": 4137 }, { "epoch": 2.7716008037508373, "grad_norm": 2.8275684672823442, "learning_rate": 1.3043854265350455e-06, "loss": 1.1487, "step": 4138 }, { "epoch": 2.772270596115204, "grad_norm": 3.8775550706414665, "learning_rate": 1.3031021230984992e-06, "loss": 1.1108, "step": 4139 }, { "epoch": 2.7729403884795714, "grad_norm": 2.630119895463535, "learning_rate": 1.301819228697926e-06, "loss": 0.984, "step": 4140 }, { "epoch": 2.7736101808439386, "grad_norm": 2.7595426379143064, "learning_rate": 1.3005367437717515e-06, "loss": 1.1113, "step": 4141 }, { "epoch": 2.7742799732083054, "grad_norm": 2.7685476094892394, "learning_rate": 1.2992546687582619e-06, "loss": 1.1245, "step": 4142 }, { "epoch": 2.7749497655726723, "grad_norm": 2.865168434846248, "learning_rate": 1.297973004095603e-06, "loss": 0.9778, "step": 4143 }, { "epoch": 2.7756195579370395, "grad_norm": 2.785298348328368, "learning_rate": 1.2966917502217769e-06, "loss": 1.1659, "step": 4144 }, { "epoch": 2.7762893503014068, "grad_norm": 3.920828123131188, "learning_rate": 1.2954109075746485e-06, "loss": 1.0334, "step": 4145 }, { "epoch": 2.7769591426657736, "grad_norm": 3.0161240238964666, "learning_rate": 1.2941304765919433e-06, "loss": 1.034, "step": 4146 }, { "epoch": 2.7776289350301404, "grad_norm": 2.457240664059968, "learning_rate": 1.2928504577112433e-06, "loss": 0.825, "step": 4147 }, { "epoch": 2.7782987273945077, "grad_norm": 2.4798275214360066, "learning_rate": 1.2915708513699923e-06, "loss": 0.9835, "step": 4148 }, { "epoch": 2.778968519758875, "grad_norm": 3.469091527106894, "learning_rate": 1.2902916580054883e-06, "loss": 1.0432, "step": 4149 }, { "epoch": 2.7796383121232418, "grad_norm": 3.155996016097194, "learning_rate": 1.2890128780548947e-06, "loss": 1.05, "step": 4150 }, { "epoch": 2.7803081044876086, "grad_norm": 2.6555351231684012, "learning_rate": 1.2877345119552292e-06, "loss": 0.7485, "step": 4151 }, { "epoch": 2.780977896851976, "grad_norm": 3.119880803057658, "learning_rate": 1.2864565601433703e-06, "loss": 1.1505, "step": 4152 }, { "epoch": 2.781647689216343, "grad_norm": 3.4014165380197436, "learning_rate": 1.285179023056051e-06, "loss": 0.9067, "step": 4153 }, { "epoch": 2.78231748158071, "grad_norm": 2.8944103572171667, "learning_rate": 1.2839019011298685e-06, "loss": 1.0447, "step": 4154 }, { "epoch": 2.782987273945077, "grad_norm": 2.5836996900448144, "learning_rate": 1.2826251948012736e-06, "loss": 1.0236, "step": 4155 }, { "epoch": 2.783657066309444, "grad_norm": 2.712617964824035, "learning_rate": 1.2813489045065783e-06, "loss": 1.0122, "step": 4156 }, { "epoch": 2.7843268586738112, "grad_norm": 3.057392888376926, "learning_rate": 1.2800730306819473e-06, "loss": 1.0504, "step": 4157 }, { "epoch": 2.784996651038178, "grad_norm": 3.231158614683929, "learning_rate": 1.2787975737634093e-06, "loss": 1.1358, "step": 4158 }, { "epoch": 2.7856664434025453, "grad_norm": 2.489514773274685, "learning_rate": 1.2775225341868466e-06, "loss": 0.9515, "step": 4159 }, { "epoch": 2.786336235766912, "grad_norm": 3.1076946677201898, "learning_rate": 1.276247912388001e-06, "loss": 0.9719, "step": 4160 }, { "epoch": 2.7870060281312794, "grad_norm": 2.559997463185347, "learning_rate": 1.2749737088024671e-06, "loss": 1.0936, "step": 4161 }, { "epoch": 2.787675820495646, "grad_norm": 3.4486966814200883, "learning_rate": 1.273699923865703e-06, "loss": 0.9729, "step": 4162 }, { "epoch": 2.7883456128600135, "grad_norm": 2.5858132119880044, "learning_rate": 1.2724265580130196e-06, "loss": 0.8799, "step": 4163 }, { "epoch": 2.7890154052243803, "grad_norm": 3.438417536766079, "learning_rate": 1.2711536116795866e-06, "loss": 0.928, "step": 4164 }, { "epoch": 2.7896851975887476, "grad_norm": 3.503726465407226, "learning_rate": 1.2698810853004263e-06, "loss": 1.1633, "step": 4165 }, { "epoch": 2.7903549899531144, "grad_norm": 4.26914727708459, "learning_rate": 1.2686089793104235e-06, "loss": 0.967, "step": 4166 }, { "epoch": 2.7910247823174816, "grad_norm": 3.47354925770082, "learning_rate": 1.2673372941443148e-06, "loss": 0.9712, "step": 4167 }, { "epoch": 2.7916945746818485, "grad_norm": 4.042246702360828, "learning_rate": 1.2660660302366954e-06, "loss": 0.8776, "step": 4168 }, { "epoch": 2.7923643670462157, "grad_norm": 2.7691251990984207, "learning_rate": 1.2647951880220147e-06, "loss": 1.1297, "step": 4169 }, { "epoch": 2.793034159410583, "grad_norm": 2.732295498026341, "learning_rate": 1.2635247679345797e-06, "loss": 1.0941, "step": 4170 }, { "epoch": 2.79370395177495, "grad_norm": 5.323281396661333, "learning_rate": 1.2622547704085513e-06, "loss": 0.9847, "step": 4171 }, { "epoch": 2.7943737441393166, "grad_norm": 2.7969804514028995, "learning_rate": 1.2609851958779484e-06, "loss": 1.0903, "step": 4172 }, { "epoch": 2.795043536503684, "grad_norm": 2.641506039612876, "learning_rate": 1.2597160447766432e-06, "loss": 1.088, "step": 4173 }, { "epoch": 2.795713328868051, "grad_norm": 2.9448989220992132, "learning_rate": 1.2584473175383639e-06, "loss": 0.977, "step": 4174 }, { "epoch": 2.796383121232418, "grad_norm": 2.895087472907584, "learning_rate": 1.2571790145966944e-06, "loss": 1.1696, "step": 4175 }, { "epoch": 2.7970529135967848, "grad_norm": 2.5529734953287724, "learning_rate": 1.2559111363850723e-06, "loss": 0.9743, "step": 4176 }, { "epoch": 2.797722705961152, "grad_norm": 4.01610704343558, "learning_rate": 1.254643683336792e-06, "loss": 1.2126, "step": 4177 }, { "epoch": 2.7983924983255193, "grad_norm": 2.6531706413555227, "learning_rate": 1.2533766558850008e-06, "loss": 1.0961, "step": 4178 }, { "epoch": 2.799062290689886, "grad_norm": 2.5481583070651737, "learning_rate": 1.252110054462701e-06, "loss": 1.0031, "step": 4179 }, { "epoch": 2.799732083054253, "grad_norm": 3.0514347584599713, "learning_rate": 1.25084387950275e-06, "loss": 1.1332, "step": 4180 }, { "epoch": 2.80040187541862, "grad_norm": 3.0111680322247123, "learning_rate": 1.249578131437859e-06, "loss": 0.9032, "step": 4181 }, { "epoch": 2.8010716677829874, "grad_norm": 2.881833923865015, "learning_rate": 1.2483128107005929e-06, "loss": 1.1279, "step": 4182 }, { "epoch": 2.8017414601473543, "grad_norm": 3.066619776271366, "learning_rate": 1.2470479177233714e-06, "loss": 1.0647, "step": 4183 }, { "epoch": 2.8024112525117215, "grad_norm": 3.112886495348071, "learning_rate": 1.2457834529384675e-06, "loss": 0.9693, "step": 4184 }, { "epoch": 2.8030810448760883, "grad_norm": 2.5389697332944055, "learning_rate": 1.2445194167780076e-06, "loss": 0.973, "step": 4185 }, { "epoch": 2.8037508372404556, "grad_norm": 2.460210690893993, "learning_rate": 1.2432558096739722e-06, "loss": 0.8132, "step": 4186 }, { "epoch": 2.8044206296048224, "grad_norm": 3.9908647338429155, "learning_rate": 1.2419926320581949e-06, "loss": 1.1092, "step": 4187 }, { "epoch": 2.8050904219691897, "grad_norm": 5.916268969112664, "learning_rate": 1.2407298843623628e-06, "loss": 0.9285, "step": 4188 }, { "epoch": 2.8057602143335565, "grad_norm": 3.5431692455277273, "learning_rate": 1.2394675670180157e-06, "loss": 1.1646, "step": 4189 }, { "epoch": 2.8064300066979238, "grad_norm": 2.6485543448595092, "learning_rate": 1.2382056804565462e-06, "loss": 0.9878, "step": 4190 }, { "epoch": 2.8070997990622906, "grad_norm": 3.612604454032307, "learning_rate": 1.2369442251092001e-06, "loss": 0.766, "step": 4191 }, { "epoch": 2.807769591426658, "grad_norm": 2.535933996462812, "learning_rate": 1.235683201407076e-06, "loss": 1.0253, "step": 4192 }, { "epoch": 2.8084393837910246, "grad_norm": 3.0085568807134933, "learning_rate": 1.2344226097811244e-06, "loss": 1.2513, "step": 4193 }, { "epoch": 2.809109176155392, "grad_norm": 2.985223280324718, "learning_rate": 1.2331624506621482e-06, "loss": 1.0797, "step": 4194 }, { "epoch": 2.8097789685197587, "grad_norm": 2.722139379865493, "learning_rate": 1.231902724480803e-06, "loss": 0.9098, "step": 4195 }, { "epoch": 2.810448760884126, "grad_norm": 2.7762300430246607, "learning_rate": 1.2306434316675958e-06, "loss": 0.8958, "step": 4196 }, { "epoch": 2.811118553248493, "grad_norm": 8.975254967015388, "learning_rate": 1.2293845726528864e-06, "loss": 0.9319, "step": 4197 }, { "epoch": 2.81178834561286, "grad_norm": 2.920410536130924, "learning_rate": 1.2281261478668855e-06, "loss": 0.9821, "step": 4198 }, { "epoch": 2.8124581379772273, "grad_norm": 2.5092351404951314, "learning_rate": 1.2268681577396555e-06, "loss": 1.0196, "step": 4199 }, { "epoch": 2.813127930341594, "grad_norm": 2.587012389635548, "learning_rate": 1.2256106027011098e-06, "loss": 1.0648, "step": 4200 }, { "epoch": 2.813797722705961, "grad_norm": 3.8708802709225663, "learning_rate": 1.2243534831810165e-06, "loss": 1.2235, "step": 4201 }, { "epoch": 2.814467515070328, "grad_norm": 4.9292518674054815, "learning_rate": 1.2230967996089893e-06, "loss": 0.9094, "step": 4202 }, { "epoch": 2.8151373074346955, "grad_norm": 4.313098638416087, "learning_rate": 1.2218405524144967e-06, "loss": 1.0914, "step": 4203 }, { "epoch": 2.8158070997990623, "grad_norm": 2.5329845989711925, "learning_rate": 1.2205847420268557e-06, "loss": 1.0856, "step": 4204 }, { "epoch": 2.816476892163429, "grad_norm": 4.214850414366911, "learning_rate": 1.219329368875239e-06, "loss": 0.8991, "step": 4205 }, { "epoch": 2.8171466845277964, "grad_norm": 3.270473143242572, "learning_rate": 1.2180744333886635e-06, "loss": 0.9771, "step": 4206 }, { "epoch": 2.8178164768921636, "grad_norm": 3.5234774856270965, "learning_rate": 1.2168199359959995e-06, "loss": 1.0434, "step": 4207 }, { "epoch": 2.8184862692565305, "grad_norm": 2.6974747572809776, "learning_rate": 1.2155658771259674e-06, "loss": 0.9542, "step": 4208 }, { "epoch": 2.8191560616208973, "grad_norm": 2.469504689598132, "learning_rate": 1.2143122572071398e-06, "loss": 1.1531, "step": 4209 }, { "epoch": 2.8198258539852645, "grad_norm": 2.68782022588122, "learning_rate": 1.2130590766679345e-06, "loss": 1.0616, "step": 4210 }, { "epoch": 2.820495646349632, "grad_norm": 3.833580765355872, "learning_rate": 1.2118063359366236e-06, "loss": 1.2401, "step": 4211 }, { "epoch": 2.8211654387139986, "grad_norm": 3.5541943139592567, "learning_rate": 1.2105540354413253e-06, "loss": 0.931, "step": 4212 }, { "epoch": 2.821835231078366, "grad_norm": 2.534712056028122, "learning_rate": 1.209302175610013e-06, "loss": 1.099, "step": 4213 }, { "epoch": 2.8225050234427327, "grad_norm": 2.737683052644773, "learning_rate": 1.208050756870502e-06, "loss": 0.9158, "step": 4214 }, { "epoch": 2.8231748158071, "grad_norm": 2.9629047223431595, "learning_rate": 1.2067997796504622e-06, "loss": 1.0179, "step": 4215 }, { "epoch": 2.8238446081714668, "grad_norm": 2.9540608868080795, "learning_rate": 1.2055492443774093e-06, "loss": 0.9994, "step": 4216 }, { "epoch": 2.824514400535834, "grad_norm": 2.7900150688279637, "learning_rate": 1.2042991514787128e-06, "loss": 1.0703, "step": 4217 }, { "epoch": 2.825184192900201, "grad_norm": 2.5647256898159814, "learning_rate": 1.203049501381587e-06, "loss": 0.9823, "step": 4218 }, { "epoch": 2.825853985264568, "grad_norm": 2.327074624882573, "learning_rate": 1.201800294513094e-06, "loss": 0.8036, "step": 4219 }, { "epoch": 2.826523777628935, "grad_norm": 5.529031100743174, "learning_rate": 1.2005515313001462e-06, "loss": 1.1039, "step": 4220 }, { "epoch": 2.827193569993302, "grad_norm": 2.930074299085516, "learning_rate": 1.1993032121695064e-06, "loss": 1.0014, "step": 4221 }, { "epoch": 2.827863362357669, "grad_norm": 3.264280791704362, "learning_rate": 1.1980553375477835e-06, "loss": 0.9363, "step": 4222 }, { "epoch": 2.8285331547220363, "grad_norm": 2.7011034555966593, "learning_rate": 1.1968079078614328e-06, "loss": 0.912, "step": 4223 }, { "epoch": 2.829202947086403, "grad_norm": 2.8014168921873455, "learning_rate": 1.1955609235367595e-06, "loss": 1.153, "step": 4224 }, { "epoch": 2.8298727394507703, "grad_norm": 3.4146432765453203, "learning_rate": 1.1943143849999184e-06, "loss": 1.1476, "step": 4225 }, { "epoch": 2.830542531815137, "grad_norm": 2.804607672800249, "learning_rate": 1.1930682926769094e-06, "loss": 0.9521, "step": 4226 }, { "epoch": 2.8312123241795044, "grad_norm": 2.7905566684674574, "learning_rate": 1.1918226469935785e-06, "loss": 0.8319, "step": 4227 }, { "epoch": 2.8318821165438717, "grad_norm": 3.0679331600350963, "learning_rate": 1.1905774483756233e-06, "loss": 1.0583, "step": 4228 }, { "epoch": 2.8325519089082385, "grad_norm": 2.602919363678856, "learning_rate": 1.1893326972485854e-06, "loss": 0.8016, "step": 4229 }, { "epoch": 2.8332217012726053, "grad_norm": 3.107272742698137, "learning_rate": 1.1880883940378562e-06, "loss": 1.1932, "step": 4230 }, { "epoch": 2.8338914936369726, "grad_norm": 2.7433437400807725, "learning_rate": 1.1868445391686687e-06, "loss": 0.9971, "step": 4231 }, { "epoch": 2.83456128600134, "grad_norm": 2.8247956645686174, "learning_rate": 1.1856011330661093e-06, "loss": 0.9974, "step": 4232 }, { "epoch": 2.8352310783657066, "grad_norm": 2.8340900280447707, "learning_rate": 1.1843581761551071e-06, "loss": 0.9496, "step": 4233 }, { "epoch": 2.8359008707300735, "grad_norm": 3.01014904758434, "learning_rate": 1.1831156688604394e-06, "loss": 0.8403, "step": 4234 }, { "epoch": 2.8365706630944407, "grad_norm": 2.9236001883735265, "learning_rate": 1.1818736116067267e-06, "loss": 1.0999, "step": 4235 }, { "epoch": 2.837240455458808, "grad_norm": 2.77807900454132, "learning_rate": 1.1806320048184405e-06, "loss": 1.033, "step": 4236 }, { "epoch": 2.837910247823175, "grad_norm": 2.430957707470538, "learning_rate": 1.1793908489198949e-06, "loss": 1.0502, "step": 4237 }, { "epoch": 2.8385800401875416, "grad_norm": 3.1477493702965287, "learning_rate": 1.178150144335252e-06, "loss": 1.126, "step": 4238 }, { "epoch": 2.839249832551909, "grad_norm": 3.026237063604038, "learning_rate": 1.1769098914885156e-06, "loss": 0.8741, "step": 4239 }, { "epoch": 2.839919624916276, "grad_norm": 2.8013263277701164, "learning_rate": 1.175670090803541e-06, "loss": 0.9938, "step": 4240 }, { "epoch": 2.840589417280643, "grad_norm": 3.165813416156778, "learning_rate": 1.1744307427040253e-06, "loss": 1.228, "step": 4241 }, { "epoch": 2.8412592096450098, "grad_norm": 3.218464455726463, "learning_rate": 1.1731918476135123e-06, "loss": 1.1194, "step": 4242 }, { "epoch": 2.841929002009377, "grad_norm": 2.5313808648028435, "learning_rate": 1.1719534059553878e-06, "loss": 1.0307, "step": 4243 }, { "epoch": 2.8425987943737443, "grad_norm": 2.8793380078954796, "learning_rate": 1.1707154181528883e-06, "loss": 0.9849, "step": 4244 }, { "epoch": 2.843268586738111, "grad_norm": 3.8809812855385415, "learning_rate": 1.1694778846290905e-06, "loss": 1.0273, "step": 4245 }, { "epoch": 2.8439383791024784, "grad_norm": 3.146026599822795, "learning_rate": 1.168240805806919e-06, "loss": 1.1881, "step": 4246 }, { "epoch": 2.844608171466845, "grad_norm": 3.245965574281286, "learning_rate": 1.1670041821091382e-06, "loss": 0.9042, "step": 4247 }, { "epoch": 2.8452779638312125, "grad_norm": 2.7282621544999888, "learning_rate": 1.1657680139583636e-06, "loss": 1.0903, "step": 4248 }, { "epoch": 2.8459477561955793, "grad_norm": 2.6595790699458615, "learning_rate": 1.1645323017770504e-06, "loss": 1.1906, "step": 4249 }, { "epoch": 2.8466175485599465, "grad_norm": 2.6179481494495014, "learning_rate": 1.1632970459875001e-06, "loss": 1.0483, "step": 4250 }, { "epoch": 2.8472873409243133, "grad_norm": 3.0653392921493725, "learning_rate": 1.162062247011855e-06, "loss": 1.0851, "step": 4251 }, { "epoch": 2.8479571332886806, "grad_norm": 2.6087508691289365, "learning_rate": 1.1608279052721058e-06, "loss": 1.0738, "step": 4252 }, { "epoch": 2.8486269256530474, "grad_norm": 2.706743080758572, "learning_rate": 1.1595940211900845e-06, "loss": 1.0289, "step": 4253 }, { "epoch": 2.8492967180174147, "grad_norm": 2.4585903717939748, "learning_rate": 1.1583605951874663e-06, "loss": 0.917, "step": 4254 }, { "epoch": 2.8499665103817815, "grad_norm": 3.4686997505172346, "learning_rate": 1.1571276276857713e-06, "loss": 1.1688, "step": 4255 }, { "epoch": 2.8506363027461488, "grad_norm": 4.143567246816746, "learning_rate": 1.1558951191063617e-06, "loss": 0.8731, "step": 4256 }, { "epoch": 2.8513060951105156, "grad_norm": 2.879608006365541, "learning_rate": 1.1546630698704434e-06, "loss": 1.1298, "step": 4257 }, { "epoch": 2.851975887474883, "grad_norm": 2.7323523826567597, "learning_rate": 1.153431480399065e-06, "loss": 1.059, "step": 4258 }, { "epoch": 2.8526456798392497, "grad_norm": 2.5500576567962243, "learning_rate": 1.1522003511131188e-06, "loss": 0.8747, "step": 4259 }, { "epoch": 2.853315472203617, "grad_norm": 3.055688726067354, "learning_rate": 1.1509696824333387e-06, "loss": 0.9548, "step": 4260 }, { "epoch": 2.853985264567984, "grad_norm": 2.950297560450078, "learning_rate": 1.149739474780302e-06, "loss": 1.0486, "step": 4261 }, { "epoch": 2.854655056932351, "grad_norm": 3.3529809113356954, "learning_rate": 1.1485097285744281e-06, "loss": 0.8342, "step": 4262 }, { "epoch": 2.855324849296718, "grad_norm": 2.795753993701898, "learning_rate": 1.1472804442359785e-06, "loss": 1.123, "step": 4263 }, { "epoch": 2.855994641661085, "grad_norm": 2.7329443918734557, "learning_rate": 1.1460516221850578e-06, "loss": 0.9563, "step": 4264 }, { "epoch": 2.8566644340254523, "grad_norm": 3.3748565866095497, "learning_rate": 1.1448232628416116e-06, "loss": 1.1109, "step": 4265 }, { "epoch": 2.857334226389819, "grad_norm": 2.793910048206842, "learning_rate": 1.1435953666254278e-06, "loss": 1.1382, "step": 4266 }, { "epoch": 2.858004018754186, "grad_norm": 2.871328420675737, "learning_rate": 1.1423679339561358e-06, "loss": 1.1018, "step": 4267 }, { "epoch": 2.8586738111185532, "grad_norm": 2.6603452363543285, "learning_rate": 1.1411409652532068e-06, "loss": 0.9995, "step": 4268 }, { "epoch": 2.8593436034829205, "grad_norm": 3.2410122223755033, "learning_rate": 1.1399144609359532e-06, "loss": 1.1133, "step": 4269 }, { "epoch": 2.8600133958472873, "grad_norm": 2.693322049592203, "learning_rate": 1.1386884214235294e-06, "loss": 0.9214, "step": 4270 }, { "epoch": 2.860683188211654, "grad_norm": 2.7155227830192517, "learning_rate": 1.1374628471349304e-06, "loss": 1.0955, "step": 4271 }, { "epoch": 2.8613529805760214, "grad_norm": 2.7102938985950495, "learning_rate": 1.1362377384889914e-06, "loss": 1.0574, "step": 4272 }, { "epoch": 2.8620227729403886, "grad_norm": 2.7178236118960926, "learning_rate": 1.13501309590439e-06, "loss": 0.9266, "step": 4273 }, { "epoch": 2.8626925653047555, "grad_norm": 3.8588940651657326, "learning_rate": 1.1337889197996438e-06, "loss": 1.0759, "step": 4274 }, { "epoch": 2.8633623576691227, "grad_norm": 2.7643777819030277, "learning_rate": 1.1325652105931113e-06, "loss": 1.0924, "step": 4275 }, { "epoch": 2.8640321500334895, "grad_norm": 3.809273480102588, "learning_rate": 1.1313419687029902e-06, "loss": 1.0751, "step": 4276 }, { "epoch": 2.864701942397857, "grad_norm": 3.349605052721512, "learning_rate": 1.1301191945473202e-06, "loss": 1.0331, "step": 4277 }, { "epoch": 2.8653717347622236, "grad_norm": 3.028053399355799, "learning_rate": 1.1288968885439805e-06, "loss": 0.923, "step": 4278 }, { "epoch": 2.866041527126591, "grad_norm": 3.023008212066739, "learning_rate": 1.1276750511106899e-06, "loss": 0.7815, "step": 4279 }, { "epoch": 2.8667113194909577, "grad_norm": 3.6203151824189885, "learning_rate": 1.1264536826650075e-06, "loss": 0.9593, "step": 4280 }, { "epoch": 2.867381111855325, "grad_norm": 3.530814486747634, "learning_rate": 1.1252327836243317e-06, "loss": 1.0042, "step": 4281 }, { "epoch": 2.8680509042196918, "grad_norm": 2.99434833605666, "learning_rate": 1.1240123544059015e-06, "loss": 0.8329, "step": 4282 }, { "epoch": 2.868720696584059, "grad_norm": 3.337542286373539, "learning_rate": 1.1227923954267938e-06, "loss": 1.039, "step": 4283 }, { "epoch": 2.869390488948426, "grad_norm": 2.7994531162069674, "learning_rate": 1.1215729071039264e-06, "loss": 1.1289, "step": 4284 }, { "epoch": 2.870060281312793, "grad_norm": 4.2748965318726695, "learning_rate": 1.120353889854055e-06, "loss": 0.7895, "step": 4285 }, { "epoch": 2.87073007367716, "grad_norm": 2.5168034576940164, "learning_rate": 1.119135344093774e-06, "loss": 1.1394, "step": 4286 }, { "epoch": 2.871399866041527, "grad_norm": 2.7417668650708453, "learning_rate": 1.1179172702395203e-06, "loss": 1.1168, "step": 4287 }, { "epoch": 2.872069658405894, "grad_norm": 4.220051521405967, "learning_rate": 1.1166996687075638e-06, "loss": 1.049, "step": 4288 }, { "epoch": 2.8727394507702613, "grad_norm": 3.0455936782801576, "learning_rate": 1.1154825399140168e-06, "loss": 0.9784, "step": 4289 }, { "epoch": 2.8734092431346285, "grad_norm": 3.0161395550630017, "learning_rate": 1.1142658842748285e-06, "loss": 1.0445, "step": 4290 }, { "epoch": 2.8740790354989953, "grad_norm": 3.167574717805517, "learning_rate": 1.1130497022057884e-06, "loss": 0.9312, "step": 4291 }, { "epoch": 2.874748827863362, "grad_norm": 2.88445322229514, "learning_rate": 1.111833994122523e-06, "loss": 0.9669, "step": 4292 }, { "epoch": 2.8754186202277294, "grad_norm": 2.8931233873450957, "learning_rate": 1.110618760440495e-06, "loss": 1.0134, "step": 4293 }, { "epoch": 2.8760884125920967, "grad_norm": 2.548059676453451, "learning_rate": 1.1094040015750063e-06, "loss": 0.8469, "step": 4294 }, { "epoch": 2.8767582049564635, "grad_norm": 2.307366495968259, "learning_rate": 1.1081897179411988e-06, "loss": 0.9923, "step": 4295 }, { "epoch": 2.8774279973208303, "grad_norm": 2.452001603582584, "learning_rate": 1.1069759099540497e-06, "loss": 1.0907, "step": 4296 }, { "epoch": 2.8780977896851976, "grad_norm": 3.035857080712431, "learning_rate": 1.1057625780283727e-06, "loss": 1.0471, "step": 4297 }, { "epoch": 2.878767582049565, "grad_norm": 2.4325506242195623, "learning_rate": 1.1045497225788192e-06, "loss": 0.8708, "step": 4298 }, { "epoch": 2.8794373744139317, "grad_norm": 2.6479916196624416, "learning_rate": 1.1033373440198814e-06, "loss": 0.9457, "step": 4299 }, { "epoch": 2.8801071667782985, "grad_norm": 12.907466176837787, "learning_rate": 1.102125442765885e-06, "loss": 1.103, "step": 4300 }, { "epoch": 2.8807769591426657, "grad_norm": 3.171000265305563, "learning_rate": 1.1009140192309919e-06, "loss": 1.0358, "step": 4301 }, { "epoch": 2.881446751507033, "grad_norm": 2.695070372188596, "learning_rate": 1.099703073829202e-06, "loss": 1.0253, "step": 4302 }, { "epoch": 2.8821165438714, "grad_norm": 2.6524420083152176, "learning_rate": 1.0984926069743537e-06, "loss": 1.0948, "step": 4303 }, { "epoch": 2.882786336235767, "grad_norm": 2.9161157993908318, "learning_rate": 1.0972826190801206e-06, "loss": 1.0807, "step": 4304 }, { "epoch": 2.883456128600134, "grad_norm": 3.470023152306496, "learning_rate": 1.0960731105600097e-06, "loss": 0.923, "step": 4305 }, { "epoch": 2.884125920964501, "grad_norm": 3.1094771848684872, "learning_rate": 1.0948640818273673e-06, "loss": 1.1751, "step": 4306 }, { "epoch": 2.884795713328868, "grad_norm": 2.7463507403626974, "learning_rate": 1.0936555332953762e-06, "loss": 1.072, "step": 4307 }, { "epoch": 2.8854655056932352, "grad_norm": 2.6579738548092586, "learning_rate": 1.0924474653770544e-06, "loss": 1.0051, "step": 4308 }, { "epoch": 2.886135298057602, "grad_norm": 3.01881450533275, "learning_rate": 1.0912398784852532e-06, "loss": 0.8938, "step": 4309 }, { "epoch": 2.8868050904219693, "grad_norm": 3.5407549907300617, "learning_rate": 1.0900327730326615e-06, "loss": 0.8902, "step": 4310 }, { "epoch": 2.887474882786336, "grad_norm": 2.7501112354883257, "learning_rate": 1.0888261494318057e-06, "loss": 1.1049, "step": 4311 }, { "epoch": 2.8881446751507034, "grad_norm": 2.7301629193155206, "learning_rate": 1.0876200080950453e-06, "loss": 0.9658, "step": 4312 }, { "epoch": 2.88881446751507, "grad_norm": 3.093992791536829, "learning_rate": 1.0864143494345734e-06, "loss": 1.0688, "step": 4313 }, { "epoch": 2.8894842598794375, "grad_norm": 2.791923223751049, "learning_rate": 1.0852091738624197e-06, "loss": 1.1432, "step": 4314 }, { "epoch": 2.8901540522438043, "grad_norm": 2.600963267684593, "learning_rate": 1.0840044817904513e-06, "loss": 1.0452, "step": 4315 }, { "epoch": 2.8908238446081715, "grad_norm": 3.2758167760171655, "learning_rate": 1.0828002736303677e-06, "loss": 0.8054, "step": 4316 }, { "epoch": 2.8914936369725384, "grad_norm": 2.633190712192616, "learning_rate": 1.0815965497937006e-06, "loss": 1.1247, "step": 4317 }, { "epoch": 2.8921634293369056, "grad_norm": 8.671931577688126, "learning_rate": 1.080393310691821e-06, "loss": 0.941, "step": 4318 }, { "epoch": 2.8928332217012724, "grad_norm": 3.1085825429742036, "learning_rate": 1.0791905567359313e-06, "loss": 0.9235, "step": 4319 }, { "epoch": 2.8935030140656397, "grad_norm": 3.0509365029372724, "learning_rate": 1.0779882883370695e-06, "loss": 1.0325, "step": 4320 }, { "epoch": 2.8941728064300065, "grad_norm": 2.8939866604348317, "learning_rate": 1.0767865059061045e-06, "loss": 1.1007, "step": 4321 }, { "epoch": 2.894842598794374, "grad_norm": 2.512019981488529, "learning_rate": 1.0755852098537441e-06, "loss": 0.9305, "step": 4322 }, { "epoch": 2.895512391158741, "grad_norm": 3.2051896198290897, "learning_rate": 1.0743844005905266e-06, "loss": 0.9451, "step": 4323 }, { "epoch": 2.896182183523108, "grad_norm": 3.8497645162301093, "learning_rate": 1.0731840785268246e-06, "loss": 0.8736, "step": 4324 }, { "epoch": 2.8968519758874747, "grad_norm": 2.620364250841252, "learning_rate": 1.0719842440728439e-06, "loss": 0.8001, "step": 4325 }, { "epoch": 2.897521768251842, "grad_norm": 2.6268925986401466, "learning_rate": 1.0707848976386249e-06, "loss": 1.0525, "step": 4326 }, { "epoch": 2.898191560616209, "grad_norm": 3.1228844295665255, "learning_rate": 1.0695860396340393e-06, "loss": 1.1084, "step": 4327 }, { "epoch": 2.898861352980576, "grad_norm": 2.4787596816867286, "learning_rate": 1.0683876704687937e-06, "loss": 1.1482, "step": 4328 }, { "epoch": 2.899531145344943, "grad_norm": 5.0076899506740755, "learning_rate": 1.0671897905524268e-06, "loss": 0.8922, "step": 4329 }, { "epoch": 2.90020093770931, "grad_norm": 8.86715534158164, "learning_rate": 1.0659924002943099e-06, "loss": 1.1253, "step": 4330 }, { "epoch": 2.9008707300736774, "grad_norm": 2.7558926123465333, "learning_rate": 1.0647955001036473e-06, "loss": 1.0745, "step": 4331 }, { "epoch": 2.901540522438044, "grad_norm": 2.7883244189764853, "learning_rate": 1.0635990903894759e-06, "loss": 0.955, "step": 4332 }, { "epoch": 2.902210314802411, "grad_norm": 2.6669677446939355, "learning_rate": 1.0624031715606648e-06, "loss": 0.9546, "step": 4333 }, { "epoch": 2.9028801071667782, "grad_norm": 2.8499499857457677, "learning_rate": 1.0612077440259152e-06, "loss": 0.9624, "step": 4334 }, { "epoch": 2.9035498995311455, "grad_norm": 16.491956159320857, "learning_rate": 1.060012808193761e-06, "loss": 0.9457, "step": 4335 }, { "epoch": 2.9042196918955123, "grad_norm": 2.649600523218373, "learning_rate": 1.0588183644725676e-06, "loss": 1.0387, "step": 4336 }, { "epoch": 2.9048894842598796, "grad_norm": 3.1117069464818377, "learning_rate": 1.057624413270532e-06, "loss": 1.0682, "step": 4337 }, { "epoch": 2.9055592766242464, "grad_norm": 5.011464260815031, "learning_rate": 1.056430954995683e-06, "loss": 0.9766, "step": 4338 }, { "epoch": 2.9062290689886137, "grad_norm": 2.8206667041733384, "learning_rate": 1.0552379900558815e-06, "loss": 1.0117, "step": 4339 }, { "epoch": 2.9068988613529805, "grad_norm": 2.879129544402116, "learning_rate": 1.0540455188588191e-06, "loss": 0.9631, "step": 4340 }, { "epoch": 2.9075686537173477, "grad_norm": 3.3082793843259504, "learning_rate": 1.0528535418120192e-06, "loss": 0.8792, "step": 4341 }, { "epoch": 2.9082384460817146, "grad_norm": 2.4845619857052545, "learning_rate": 1.0516620593228364e-06, "loss": 1.0356, "step": 4342 }, { "epoch": 2.908908238446082, "grad_norm": 2.6594679135977284, "learning_rate": 1.0504710717984551e-06, "loss": 1.0113, "step": 4343 }, { "epoch": 2.9095780308104486, "grad_norm": 3.3225185495342333, "learning_rate": 1.049280579645892e-06, "loss": 1.2931, "step": 4344 }, { "epoch": 2.910247823174816, "grad_norm": 2.7490178925353588, "learning_rate": 1.0480905832719944e-06, "loss": 1.1972, "step": 4345 }, { "epoch": 2.9109176155391827, "grad_norm": 2.9859450833090264, "learning_rate": 1.0469010830834386e-06, "loss": 1.0177, "step": 4346 }, { "epoch": 2.91158740790355, "grad_norm": 3.0951398492622886, "learning_rate": 1.0457120794867334e-06, "loss": 0.9744, "step": 4347 }, { "epoch": 2.912257200267917, "grad_norm": 2.7401756782097335, "learning_rate": 1.0445235728882164e-06, "loss": 0.968, "step": 4348 }, { "epoch": 2.912926992632284, "grad_norm": 2.6427297839412276, "learning_rate": 1.043335563694056e-06, "loss": 0.9286, "step": 4349 }, { "epoch": 2.913596784996651, "grad_norm": 3.316215508325491, "learning_rate": 1.0421480523102506e-06, "loss": 1.1323, "step": 4350 }, { "epoch": 2.914266577361018, "grad_norm": 3.215591427362974, "learning_rate": 1.0409610391426284e-06, "loss": 1.0152, "step": 4351 }, { "epoch": 2.9149363697253854, "grad_norm": 3.6473584719041097, "learning_rate": 1.039774524596847e-06, "loss": 1.0665, "step": 4352 }, { "epoch": 2.915606162089752, "grad_norm": 2.8502451726592093, "learning_rate": 1.0385885090783942e-06, "loss": 0.9547, "step": 4353 }, { "epoch": 2.916275954454119, "grad_norm": 2.7686267076513342, "learning_rate": 1.037402992992587e-06, "loss": 1.0921, "step": 4354 }, { "epoch": 2.9169457468184863, "grad_norm": 2.816773821632984, "learning_rate": 1.0362179767445715e-06, "loss": 1.1879, "step": 4355 }, { "epoch": 2.9176155391828535, "grad_norm": 2.513855705248693, "learning_rate": 1.035033460739323e-06, "loss": 0.9451, "step": 4356 }, { "epoch": 2.9182853315472204, "grad_norm": 2.8061990113597193, "learning_rate": 1.0338494453816466e-06, "loss": 0.9842, "step": 4357 }, { "epoch": 2.918955123911587, "grad_norm": 2.7181174345412966, "learning_rate": 1.0326659310761752e-06, "loss": 0.9062, "step": 4358 }, { "epoch": 2.9196249162759544, "grad_norm": 2.720587707729937, "learning_rate": 1.031482918227371e-06, "loss": 1.0673, "step": 4359 }, { "epoch": 2.9202947086403217, "grad_norm": 2.7827322740796467, "learning_rate": 1.0303004072395252e-06, "loss": 1.2028, "step": 4360 }, { "epoch": 2.9209645010046885, "grad_norm": 2.519589750979188, "learning_rate": 1.029118398516756e-06, "loss": 1.0793, "step": 4361 }, { "epoch": 2.9216342933690553, "grad_norm": 3.292597551041079, "learning_rate": 1.0279368924630133e-06, "loss": 1.0, "step": 4362 }, { "epoch": 2.9223040857334226, "grad_norm": 3.714904765892452, "learning_rate": 1.0267558894820713e-06, "loss": 0.9561, "step": 4363 }, { "epoch": 2.92297387809779, "grad_norm": 2.681357717973602, "learning_rate": 1.0255753899775336e-06, "loss": 1.1002, "step": 4364 }, { "epoch": 2.9236436704621567, "grad_norm": 3.7281965611187653, "learning_rate": 1.0243953943528321e-06, "loss": 0.9141, "step": 4365 }, { "epoch": 2.924313462826524, "grad_norm": 2.981093764978178, "learning_rate": 1.0232159030112287e-06, "loss": 1.1111, "step": 4366 }, { "epoch": 2.9249832551908908, "grad_norm": 2.701664038685233, "learning_rate": 1.0220369163558083e-06, "loss": 1.032, "step": 4367 }, { "epoch": 2.925653047555258, "grad_norm": 3.3725076413379367, "learning_rate": 1.0208584347894868e-06, "loss": 0.9612, "step": 4368 }, { "epoch": 2.926322839919625, "grad_norm": 2.532503079418348, "learning_rate": 1.0196804587150053e-06, "loss": 0.9663, "step": 4369 }, { "epoch": 2.926992632283992, "grad_norm": 3.2572318284955917, "learning_rate": 1.018502988534936e-06, "loss": 1.0209, "step": 4370 }, { "epoch": 2.927662424648359, "grad_norm": 3.2049388501878187, "learning_rate": 1.0173260246516728e-06, "loss": 0.9653, "step": 4371 }, { "epoch": 2.928332217012726, "grad_norm": 2.909724312472035, "learning_rate": 1.0161495674674404e-06, "loss": 1.0092, "step": 4372 }, { "epoch": 2.929002009377093, "grad_norm": 2.8674086651272566, "learning_rate": 1.0149736173842884e-06, "loss": 1.0436, "step": 4373 }, { "epoch": 2.9296718017414602, "grad_norm": 3.2107657318413523, "learning_rate": 1.0137981748040963e-06, "loss": 1.0497, "step": 4374 }, { "epoch": 2.930341594105827, "grad_norm": 2.624677715606942, "learning_rate": 1.012623240128565e-06, "loss": 1.0861, "step": 4375 }, { "epoch": 2.9310113864701943, "grad_norm": 3.106857253459247, "learning_rate": 1.0114488137592252e-06, "loss": 1.0491, "step": 4376 }, { "epoch": 2.931681178834561, "grad_norm": 2.742505095177992, "learning_rate": 1.0102748960974345e-06, "loss": 0.9047, "step": 4377 }, { "epoch": 2.9323509711989284, "grad_norm": 3.3421866644407547, "learning_rate": 1.009101487544376e-06, "loss": 0.9725, "step": 4378 }, { "epoch": 2.933020763563295, "grad_norm": 3.249333551198184, "learning_rate": 1.0079285885010563e-06, "loss": 0.9104, "step": 4379 }, { "epoch": 2.9336905559276625, "grad_norm": 2.745335233293622, "learning_rate": 1.0067561993683097e-06, "loss": 1.0389, "step": 4380 }, { "epoch": 2.9343603482920297, "grad_norm": 3.1909225371107515, "learning_rate": 1.005584320546798e-06, "loss": 0.9835, "step": 4381 }, { "epoch": 2.9350301406563966, "grad_norm": 2.6475218838349743, "learning_rate": 1.0044129524370074e-06, "loss": 1.1898, "step": 4382 }, { "epoch": 2.9356999330207634, "grad_norm": 2.7434222908031405, "learning_rate": 1.0032420954392475e-06, "loss": 1.0043, "step": 4383 }, { "epoch": 2.9363697253851306, "grad_norm": 2.9858710946167277, "learning_rate": 1.0020717499536543e-06, "loss": 0.9493, "step": 4384 }, { "epoch": 2.937039517749498, "grad_norm": 2.815312037156135, "learning_rate": 1.0009019163801918e-06, "loss": 1.1287, "step": 4385 }, { "epoch": 2.9377093101138647, "grad_norm": 2.6536804673527477, "learning_rate": 9.997325951186468e-07, "loss": 1.0178, "step": 4386 }, { "epoch": 2.9383791024782315, "grad_norm": 2.499419031530806, "learning_rate": 9.98563786568629e-07, "loss": 0.9866, "step": 4387 }, { "epoch": 2.939048894842599, "grad_norm": 3.026608089705332, "learning_rate": 9.973954911295749e-07, "loss": 1.1267, "step": 4388 }, { "epoch": 2.939718687206966, "grad_norm": 2.9469173467711274, "learning_rate": 9.96227709200748e-07, "loss": 1.1998, "step": 4389 }, { "epoch": 2.940388479571333, "grad_norm": 3.1122191957017464, "learning_rate": 9.950604411812337e-07, "loss": 0.9321, "step": 4390 }, { "epoch": 2.9410582719356997, "grad_norm": 4.120629559389471, "learning_rate": 9.938936874699404e-07, "loss": 0.819, "step": 4391 }, { "epoch": 2.941728064300067, "grad_norm": 2.8888695644003466, "learning_rate": 9.927274484656018e-07, "loss": 1.0503, "step": 4392 }, { "epoch": 2.942397856664434, "grad_norm": 2.450000646002941, "learning_rate": 9.915617245667792e-07, "loss": 0.9519, "step": 4393 }, { "epoch": 2.943067649028801, "grad_norm": 2.6620568237227764, "learning_rate": 9.903965161718538e-07, "loss": 1.0805, "step": 4394 }, { "epoch": 2.943737441393168, "grad_norm": 3.0278968004688367, "learning_rate": 9.892318236790307e-07, "loss": 1.0151, "step": 4395 }, { "epoch": 2.944407233757535, "grad_norm": 3.1815638181151793, "learning_rate": 9.880676474863396e-07, "loss": 1.1601, "step": 4396 }, { "epoch": 2.9450770261219024, "grad_norm": 2.8872149674493315, "learning_rate": 9.869039879916356e-07, "loss": 1.1053, "step": 4397 }, { "epoch": 2.945746818486269, "grad_norm": 2.767964452769414, "learning_rate": 9.857408455925952e-07, "loss": 1.0079, "step": 4398 }, { "epoch": 2.9464166108506364, "grad_norm": 2.718427946152317, "learning_rate": 9.845782206867189e-07, "loss": 1.0028, "step": 4399 }, { "epoch": 2.9470864032150033, "grad_norm": 2.777749160992606, "learning_rate": 9.83416113671327e-07, "loss": 1.0192, "step": 4400 }, { "epoch": 2.9477561955793705, "grad_norm": 3.2039146099329225, "learning_rate": 9.822545249435696e-07, "loss": 1.0333, "step": 4401 }, { "epoch": 2.9484259879437373, "grad_norm": 2.731050482653438, "learning_rate": 9.810934549004142e-07, "loss": 0.8773, "step": 4402 }, { "epoch": 2.9490957803081046, "grad_norm": 3.7319861325557055, "learning_rate": 9.799329039386537e-07, "loss": 0.9773, "step": 4403 }, { "epoch": 2.9497655726724714, "grad_norm": 3.580172437458483, "learning_rate": 9.787728724549002e-07, "loss": 0.9352, "step": 4404 }, { "epoch": 2.9504353650368387, "grad_norm": 2.4170727385336064, "learning_rate": 9.776133608455931e-07, "loss": 0.987, "step": 4405 }, { "epoch": 2.9511051574012055, "grad_norm": 2.8757582084492084, "learning_rate": 9.764543695069912e-07, "loss": 0.956, "step": 4406 }, { "epoch": 2.9517749497655728, "grad_norm": 2.670526678276944, "learning_rate": 9.752958988351757e-07, "loss": 1.0357, "step": 4407 }, { "epoch": 2.9524447421299396, "grad_norm": 2.7979597880335842, "learning_rate": 9.741379492260503e-07, "loss": 0.8726, "step": 4408 }, { "epoch": 2.953114534494307, "grad_norm": 3.8085468263984192, "learning_rate": 9.729805210753404e-07, "loss": 0.9967, "step": 4409 }, { "epoch": 2.9537843268586736, "grad_norm": 3.13775785368889, "learning_rate": 9.718236147785934e-07, "loss": 0.9538, "step": 4410 }, { "epoch": 2.954454119223041, "grad_norm": 3.17287572076787, "learning_rate": 9.706672307311785e-07, "loss": 0.8953, "step": 4411 }, { "epoch": 2.9551239115874077, "grad_norm": 3.2615314109728164, "learning_rate": 9.695113693282863e-07, "loss": 1.0492, "step": 4412 }, { "epoch": 2.955793703951775, "grad_norm": 2.976768281597742, "learning_rate": 9.683560309649281e-07, "loss": 1.049, "step": 4413 }, { "epoch": 2.9564634963161422, "grad_norm": 2.7378755379501634, "learning_rate": 9.672012160359373e-07, "loss": 1.0825, "step": 4414 }, { "epoch": 2.957133288680509, "grad_norm": 2.5604776156818807, "learning_rate": 9.660469249359686e-07, "loss": 0.9717, "step": 4415 }, { "epoch": 2.957803081044876, "grad_norm": 2.6615668045384293, "learning_rate": 9.648931580594967e-07, "loss": 1.0347, "step": 4416 }, { "epoch": 2.958472873409243, "grad_norm": 2.7407406030696784, "learning_rate": 9.637399158008178e-07, "loss": 1.0237, "step": 4417 }, { "epoch": 2.9591426657736104, "grad_norm": 2.6138025712949706, "learning_rate": 9.62587198554049e-07, "loss": 0.9764, "step": 4418 }, { "epoch": 2.959812458137977, "grad_norm": 2.900978886916491, "learning_rate": 9.614350067131274e-07, "loss": 0.9919, "step": 4419 }, { "epoch": 2.960482250502344, "grad_norm": 3.3333762844554458, "learning_rate": 9.60283340671811e-07, "loss": 0.9412, "step": 4420 }, { "epoch": 2.9611520428667113, "grad_norm": 3.0207515845262054, "learning_rate": 9.591322008236779e-07, "loss": 1.1931, "step": 4421 }, { "epoch": 2.9618218352310786, "grad_norm": 3.372137031404483, "learning_rate": 9.57981587562126e-07, "loss": 0.9759, "step": 4422 }, { "epoch": 2.9624916275954454, "grad_norm": 2.6582440473750593, "learning_rate": 9.568315012803744e-07, "loss": 0.9187, "step": 4423 }, { "epoch": 2.963161419959812, "grad_norm": 2.5677370538451916, "learning_rate": 9.556819423714607e-07, "loss": 0.9914, "step": 4424 }, { "epoch": 2.9638312123241795, "grad_norm": 3.1375219411124453, "learning_rate": 9.545329112282436e-07, "loss": 0.9201, "step": 4425 }, { "epoch": 2.9645010046885467, "grad_norm": 2.7944312510599825, "learning_rate": 9.533844082434001e-07, "loss": 1.0032, "step": 4426 }, { "epoch": 2.9651707970529135, "grad_norm": 2.5410592593346686, "learning_rate": 9.522364338094278e-07, "loss": 1.0047, "step": 4427 }, { "epoch": 2.965840589417281, "grad_norm": 2.9950470637302105, "learning_rate": 9.510889883186431e-07, "loss": 1.0627, "step": 4428 }, { "epoch": 2.9665103817816476, "grad_norm": 2.628210966249791, "learning_rate": 9.499420721631819e-07, "loss": 1.049, "step": 4429 }, { "epoch": 2.967180174146015, "grad_norm": 2.822368632644211, "learning_rate": 9.487956857349989e-07, "loss": 1.1285, "step": 4430 }, { "epoch": 2.9678499665103817, "grad_norm": 3.2811318927510853, "learning_rate": 9.476498294258682e-07, "loss": 0.9553, "step": 4431 }, { "epoch": 2.968519758874749, "grad_norm": 2.9055924809114235, "learning_rate": 9.46504503627382e-07, "loss": 0.9506, "step": 4432 }, { "epoch": 2.9691895512391158, "grad_norm": 2.8943735278314775, "learning_rate": 9.453597087309521e-07, "loss": 1.0236, "step": 4433 }, { "epoch": 2.969859343603483, "grad_norm": 2.6193345635280494, "learning_rate": 9.442154451278085e-07, "loss": 0.9763, "step": 4434 }, { "epoch": 2.97052913596785, "grad_norm": 3.5971737456062938, "learning_rate": 9.430717132089995e-07, "loss": 0.9553, "step": 4435 }, { "epoch": 2.971198928332217, "grad_norm": 2.687621318703107, "learning_rate": 9.419285133653916e-07, "loss": 1.0329, "step": 4436 }, { "epoch": 2.971868720696584, "grad_norm": 2.9434949680472675, "learning_rate": 9.407858459876698e-07, "loss": 1.09, "step": 4437 }, { "epoch": 2.972538513060951, "grad_norm": 2.8681796594328954, "learning_rate": 9.396437114663368e-07, "loss": 1.0087, "step": 4438 }, { "epoch": 2.973208305425318, "grad_norm": 2.767977769591096, "learning_rate": 9.385021101917127e-07, "loss": 1.0316, "step": 4439 }, { "epoch": 2.9738780977896853, "grad_norm": 2.6919303118694535, "learning_rate": 9.373610425539386e-07, "loss": 1.0427, "step": 4440 }, { "epoch": 2.974547890154052, "grad_norm": 2.855557289125745, "learning_rate": 9.362205089429682e-07, "loss": 1.1305, "step": 4441 }, { "epoch": 2.9752176825184193, "grad_norm": 2.9758352930739878, "learning_rate": 9.35080509748576e-07, "loss": 0.9291, "step": 4442 }, { "epoch": 2.9758874748827866, "grad_norm": 2.82606696696787, "learning_rate": 9.339410453603517e-07, "loss": 1.1219, "step": 4443 }, { "epoch": 2.9765572672471534, "grad_norm": 2.97165708061833, "learning_rate": 9.328021161677067e-07, "loss": 1.0401, "step": 4444 }, { "epoch": 2.9772270596115202, "grad_norm": 3.0979994325831637, "learning_rate": 9.316637225598638e-07, "loss": 1.1299, "step": 4445 }, { "epoch": 2.9778968519758875, "grad_norm": 3.7070391299826433, "learning_rate": 9.305258649258656e-07, "loss": 1.0439, "step": 4446 }, { "epoch": 2.9785666443402548, "grad_norm": 3.1808291890188323, "learning_rate": 9.293885436545711e-07, "loss": 0.8758, "step": 4447 }, { "epoch": 2.9792364367046216, "grad_norm": 2.507300551406138, "learning_rate": 9.282517591346582e-07, "loss": 0.8795, "step": 4448 }, { "epoch": 2.9799062290689884, "grad_norm": 3.086293953792713, "learning_rate": 9.271155117546171e-07, "loss": 1.216, "step": 4449 }, { "epoch": 2.9805760214333556, "grad_norm": 3.9737536004387435, "learning_rate": 9.259798019027574e-07, "loss": 0.9968, "step": 4450 }, { "epoch": 2.981245813797723, "grad_norm": 4.3638471372810335, "learning_rate": 9.248446299672034e-07, "loss": 0.9741, "step": 4451 }, { "epoch": 2.9819156061620897, "grad_norm": 3.173273361621275, "learning_rate": 9.23709996335899e-07, "loss": 1.0608, "step": 4452 }, { "epoch": 2.9825853985264565, "grad_norm": 3.0131452374588306, "learning_rate": 9.22575901396599e-07, "loss": 0.8261, "step": 4453 }, { "epoch": 2.983255190890824, "grad_norm": 5.346605305714164, "learning_rate": 9.214423455368776e-07, "loss": 0.9534, "step": 4454 }, { "epoch": 2.983924983255191, "grad_norm": 3.8804969340245092, "learning_rate": 9.203093291441228e-07, "loss": 1.0714, "step": 4455 }, { "epoch": 2.984594775619558, "grad_norm": 5.311998099933741, "learning_rate": 9.191768526055423e-07, "loss": 0.9453, "step": 4456 }, { "epoch": 2.985264567983925, "grad_norm": 2.793851385173408, "learning_rate": 9.180449163081532e-07, "loss": 0.7802, "step": 4457 }, { "epoch": 2.985934360348292, "grad_norm": 3.422379186515471, "learning_rate": 9.169135206387919e-07, "loss": 0.9382, "step": 4458 }, { "epoch": 2.986604152712659, "grad_norm": 2.784693568571555, "learning_rate": 9.15782665984109e-07, "loss": 1.0538, "step": 4459 }, { "epoch": 2.987273945077026, "grad_norm": 2.8328112467896887, "learning_rate": 9.146523527305723e-07, "loss": 1.0065, "step": 4460 }, { "epoch": 2.9879437374413933, "grad_norm": 2.860419259567369, "learning_rate": 9.135225812644605e-07, "loss": 1.1475, "step": 4461 }, { "epoch": 2.98861352980576, "grad_norm": 3.2585812480386616, "learning_rate": 9.1239335197187e-07, "loss": 0.9345, "step": 4462 }, { "epoch": 2.9892833221701274, "grad_norm": 3.006024662860969, "learning_rate": 9.112646652387108e-07, "loss": 1.0951, "step": 4463 }, { "epoch": 2.989953114534494, "grad_norm": 3.1713908166633296, "learning_rate": 9.101365214507099e-07, "loss": 1.0181, "step": 4464 }, { "epoch": 2.9906229068988615, "grad_norm": 3.0890306648442696, "learning_rate": 9.090089209934044e-07, "loss": 0.9735, "step": 4465 }, { "epoch": 2.9912926992632283, "grad_norm": 4.472301190493141, "learning_rate": 9.078818642521487e-07, "loss": 1.1008, "step": 4466 }, { "epoch": 2.9919624916275955, "grad_norm": 2.561863189428027, "learning_rate": 9.067553516121116e-07, "loss": 1.0561, "step": 4467 }, { "epoch": 2.9926322839919623, "grad_norm": 2.603011804174436, "learning_rate": 9.056293834582749e-07, "loss": 0.9729, "step": 4468 }, { "epoch": 2.9933020763563296, "grad_norm": 2.693060189386872, "learning_rate": 9.045039601754352e-07, "loss": 0.9206, "step": 4469 }, { "epoch": 2.9939718687206964, "grad_norm": 2.807538679650497, "learning_rate": 9.033790821481997e-07, "loss": 0.9721, "step": 4470 }, { "epoch": 2.9946416610850637, "grad_norm": 2.784126700558592, "learning_rate": 9.022547497609943e-07, "loss": 1.0556, "step": 4471 }, { "epoch": 2.995311453449431, "grad_norm": 2.823295562159539, "learning_rate": 9.011309633980555e-07, "loss": 1.1759, "step": 4472 }, { "epoch": 2.9959812458137978, "grad_norm": 2.571370464768894, "learning_rate": 9.00007723443434e-07, "loss": 0.9569, "step": 4473 }, { "epoch": 2.9966510381781646, "grad_norm": 2.815519491199052, "learning_rate": 8.988850302809912e-07, "loss": 1.0229, "step": 4474 }, { "epoch": 2.997320830542532, "grad_norm": 2.654373426356516, "learning_rate": 8.977628842944064e-07, "loss": 0.985, "step": 4475 }, { "epoch": 2.997990622906899, "grad_norm": 3.035996701052248, "learning_rate": 8.966412858671683e-07, "loss": 0.9449, "step": 4476 }, { "epoch": 2.998660415271266, "grad_norm": 3.078737459420171, "learning_rate": 8.955202353825804e-07, "loss": 0.9545, "step": 4477 }, { "epoch": 2.9993302076356327, "grad_norm": 3.135286208137656, "learning_rate": 8.943997332237559e-07, "loss": 1.0867, "step": 4478 }, { "epoch": 3.0, "grad_norm": 2.6217025681313477, "learning_rate": 8.932797797736247e-07, "loss": 0.8909, "step": 4479 }, { "epoch": 3.0006697923643673, "grad_norm": 2.667132228226188, "learning_rate": 8.921603754149269e-07, "loss": 0.7286, "step": 4480 }, { "epoch": 3.001339584728734, "grad_norm": 2.934089065112888, "learning_rate": 8.910415205302161e-07, "loss": 0.9962, "step": 4481 }, { "epoch": 3.0020093770931013, "grad_norm": 2.4086245406566538, "learning_rate": 8.899232155018544e-07, "loss": 0.815, "step": 4482 }, { "epoch": 3.002679169457468, "grad_norm": 2.584204174481187, "learning_rate": 8.888054607120222e-07, "loss": 0.9237, "step": 4483 }, { "epoch": 3.0033489618218354, "grad_norm": 3.2330668885169436, "learning_rate": 8.876882565427072e-07, "loss": 0.9907, "step": 4484 }, { "epoch": 3.0040187541862022, "grad_norm": 3.521830332252582, "learning_rate": 8.86571603375711e-07, "loss": 0.697, "step": 4485 }, { "epoch": 3.0046885465505695, "grad_norm": 3.1166309426882193, "learning_rate": 8.854555015926442e-07, "loss": 0.9526, "step": 4486 }, { "epoch": 3.0053583389149363, "grad_norm": 4.069690865235747, "learning_rate": 8.843399515749332e-07, "loss": 0.9291, "step": 4487 }, { "epoch": 3.0060281312793036, "grad_norm": 2.438444951399153, "learning_rate": 8.832249537038129e-07, "loss": 0.7548, "step": 4488 }, { "epoch": 3.0066979236436704, "grad_norm": 2.9852665452676588, "learning_rate": 8.82110508360331e-07, "loss": 1.061, "step": 4489 }, { "epoch": 3.0073677160080377, "grad_norm": 2.6567976979414927, "learning_rate": 8.809966159253427e-07, "loss": 0.8962, "step": 4490 }, { "epoch": 3.0080375083724045, "grad_norm": 2.783582902153481, "learning_rate": 8.798832767795201e-07, "loss": 0.9079, "step": 4491 }, { "epoch": 3.0087073007367717, "grad_norm": 2.8259133390922444, "learning_rate": 8.787704913033424e-07, "loss": 1.0427, "step": 4492 }, { "epoch": 3.0093770931011385, "grad_norm": 2.939203337054024, "learning_rate": 8.776582598771011e-07, "loss": 0.7984, "step": 4493 }, { "epoch": 3.010046885465506, "grad_norm": 2.4868627756298114, "learning_rate": 8.765465828808953e-07, "loss": 0.9, "step": 4494 }, { "epoch": 3.0107166778298726, "grad_norm": 2.7113262709154444, "learning_rate": 8.754354606946394e-07, "loss": 0.8918, "step": 4495 }, { "epoch": 3.01138647019424, "grad_norm": 2.414546620693029, "learning_rate": 8.74324893698055e-07, "loss": 0.7388, "step": 4496 }, { "epoch": 3.0120562625586067, "grad_norm": 2.276575771149887, "learning_rate": 8.732148822706748e-07, "loss": 0.7323, "step": 4497 }, { "epoch": 3.012726054922974, "grad_norm": 2.7476507142727677, "learning_rate": 8.721054267918416e-07, "loss": 0.9088, "step": 4498 }, { "epoch": 3.013395847287341, "grad_norm": 3.0191492697685494, "learning_rate": 8.709965276407084e-07, "loss": 0.8642, "step": 4499 }, { "epoch": 3.014065639651708, "grad_norm": 3.1197013074692963, "learning_rate": 8.698881851962377e-07, "loss": 0.8687, "step": 4500 }, { "epoch": 3.014735432016075, "grad_norm": 2.8914257254963465, "learning_rate": 8.687803998372019e-07, "loss": 0.7826, "step": 4501 }, { "epoch": 3.015405224380442, "grad_norm": 2.98244171414246, "learning_rate": 8.67673171942183e-07, "loss": 0.741, "step": 4502 }, { "epoch": 3.016075016744809, "grad_norm": 3.1410899677236523, "learning_rate": 8.665665018895728e-07, "loss": 0.7782, "step": 4503 }, { "epoch": 3.016744809109176, "grad_norm": 3.03339008272557, "learning_rate": 8.65460390057572e-07, "loss": 0.9227, "step": 4504 }, { "epoch": 3.017414601473543, "grad_norm": 3.0613512967765892, "learning_rate": 8.643548368241911e-07, "loss": 0.9262, "step": 4505 }, { "epoch": 3.0180843938379103, "grad_norm": 2.7815073090738514, "learning_rate": 8.632498425672489e-07, "loss": 0.8927, "step": 4506 }, { "epoch": 3.018754186202277, "grad_norm": 2.8814851960219467, "learning_rate": 8.621454076643737e-07, "loss": 0.9607, "step": 4507 }, { "epoch": 3.0194239785666444, "grad_norm": 2.837752373226827, "learning_rate": 8.610415324930027e-07, "loss": 0.7726, "step": 4508 }, { "epoch": 3.020093770931011, "grad_norm": 3.0574928208222065, "learning_rate": 8.599382174303816e-07, "loss": 0.8917, "step": 4509 }, { "epoch": 3.0207635632953784, "grad_norm": 2.8815808202026614, "learning_rate": 8.588354628535648e-07, "loss": 0.825, "step": 4510 }, { "epoch": 3.0214333556597457, "grad_norm": 3.139990453691415, "learning_rate": 8.57733269139415e-07, "loss": 0.9437, "step": 4511 }, { "epoch": 3.0221031480241125, "grad_norm": 3.138027060942485, "learning_rate": 8.566316366646033e-07, "loss": 0.8661, "step": 4512 }, { "epoch": 3.0227729403884798, "grad_norm": 3.372047804104106, "learning_rate": 8.555305658056092e-07, "loss": 1.0189, "step": 4513 }, { "epoch": 3.0234427327528466, "grad_norm": 2.968758307396121, "learning_rate": 8.544300569387201e-07, "loss": 0.9043, "step": 4514 }, { "epoch": 3.024112525117214, "grad_norm": 2.6705184875465795, "learning_rate": 8.53330110440031e-07, "loss": 0.7232, "step": 4515 }, { "epoch": 3.0247823174815807, "grad_norm": 3.142285494996528, "learning_rate": 8.522307266854457e-07, "loss": 1.0432, "step": 4516 }, { "epoch": 3.025452109845948, "grad_norm": 2.4548227306859105, "learning_rate": 8.511319060506743e-07, "loss": 0.8666, "step": 4517 }, { "epoch": 3.0261219022103147, "grad_norm": 2.833486097107631, "learning_rate": 8.500336489112357e-07, "loss": 0.9564, "step": 4518 }, { "epoch": 3.026791694574682, "grad_norm": 2.932739564575375, "learning_rate": 8.489359556424557e-07, "loss": 0.9172, "step": 4519 }, { "epoch": 3.027461486939049, "grad_norm": 3.075654597187827, "learning_rate": 8.47838826619467e-07, "loss": 0.9398, "step": 4520 }, { "epoch": 3.028131279303416, "grad_norm": 2.9577597512354137, "learning_rate": 8.467422622172102e-07, "loss": 0.9315, "step": 4521 }, { "epoch": 3.028801071667783, "grad_norm": 2.9771940224262794, "learning_rate": 8.456462628104329e-07, "loss": 0.9497, "step": 4522 }, { "epoch": 3.02947086403215, "grad_norm": 2.7784151689681766, "learning_rate": 8.445508287736886e-07, "loss": 0.9275, "step": 4523 }, { "epoch": 3.030140656396517, "grad_norm": 2.8584399507148732, "learning_rate": 8.434559604813391e-07, "loss": 0.9235, "step": 4524 }, { "epoch": 3.0308104487608842, "grad_norm": 3.08062513523288, "learning_rate": 8.423616583075511e-07, "loss": 0.9642, "step": 4525 }, { "epoch": 3.031480241125251, "grad_norm": 3.0200414449854094, "learning_rate": 8.412679226263009e-07, "loss": 1.0315, "step": 4526 }, { "epoch": 3.0321500334896183, "grad_norm": 3.197853947500942, "learning_rate": 8.401747538113672e-07, "loss": 1.0609, "step": 4527 }, { "epoch": 3.032819825853985, "grad_norm": 2.885378116291476, "learning_rate": 8.390821522363374e-07, "loss": 0.9546, "step": 4528 }, { "epoch": 3.0334896182183524, "grad_norm": 3.044658981159335, "learning_rate": 8.379901182746036e-07, "loss": 0.8835, "step": 4529 }, { "epoch": 3.034159410582719, "grad_norm": 3.5989322099009247, "learning_rate": 8.368986522993675e-07, "loss": 0.9228, "step": 4530 }, { "epoch": 3.0348292029470865, "grad_norm": 2.771730083993289, "learning_rate": 8.358077546836319e-07, "loss": 0.9541, "step": 4531 }, { "epoch": 3.0354989953114533, "grad_norm": 3.1911874162034923, "learning_rate": 8.347174258002077e-07, "loss": 0.9576, "step": 4532 }, { "epoch": 3.0361687876758205, "grad_norm": 3.941128383654233, "learning_rate": 8.33627666021711e-07, "loss": 0.7137, "step": 4533 }, { "epoch": 3.0368385800401874, "grad_norm": 3.5699103209122343, "learning_rate": 8.325384757205659e-07, "loss": 0.7267, "step": 4534 }, { "epoch": 3.0375083724045546, "grad_norm": 3.711552268626049, "learning_rate": 8.314498552689973e-07, "loss": 0.8913, "step": 4535 }, { "epoch": 3.0381781647689214, "grad_norm": 2.992441480834446, "learning_rate": 8.303618050390383e-07, "loss": 0.9633, "step": 4536 }, { "epoch": 3.0388479571332887, "grad_norm": 2.889199421566829, "learning_rate": 8.292743254025259e-07, "loss": 0.8421, "step": 4537 }, { "epoch": 3.0395177494976555, "grad_norm": 2.6171095396974042, "learning_rate": 8.281874167311054e-07, "loss": 0.8172, "step": 4538 }, { "epoch": 3.040187541862023, "grad_norm": 3.08984264468204, "learning_rate": 8.271010793962217e-07, "loss": 0.9936, "step": 4539 }, { "epoch": 3.0408573342263896, "grad_norm": 2.7113256041397356, "learning_rate": 8.260153137691276e-07, "loss": 0.9138, "step": 4540 }, { "epoch": 3.041527126590757, "grad_norm": 2.913965893797083, "learning_rate": 8.249301202208799e-07, "loss": 0.8329, "step": 4541 }, { "epoch": 3.042196918955124, "grad_norm": 2.8690857665666676, "learning_rate": 8.238454991223407e-07, "loss": 0.9085, "step": 4542 }, { "epoch": 3.042866711319491, "grad_norm": 2.930822954844864, "learning_rate": 8.227614508441761e-07, "loss": 0.9621, "step": 4543 }, { "epoch": 3.043536503683858, "grad_norm": 3.013820795291381, "learning_rate": 8.216779757568549e-07, "loss": 0.8922, "step": 4544 }, { "epoch": 3.044206296048225, "grad_norm": 3.1737788039682684, "learning_rate": 8.205950742306506e-07, "loss": 0.901, "step": 4545 }, { "epoch": 3.0448760884125923, "grad_norm": 3.2082696230307604, "learning_rate": 8.195127466356434e-07, "loss": 0.9829, "step": 4546 }, { "epoch": 3.045545880776959, "grad_norm": 3.173184430718217, "learning_rate": 8.184309933417147e-07, "loss": 0.7912, "step": 4547 }, { "epoch": 3.0462156731413264, "grad_norm": 2.7492701890345557, "learning_rate": 8.173498147185491e-07, "loss": 0.8286, "step": 4548 }, { "epoch": 3.046885465505693, "grad_norm": 2.687096666930007, "learning_rate": 8.162692111356357e-07, "loss": 0.9514, "step": 4549 }, { "epoch": 3.0475552578700604, "grad_norm": 3.1717447556422953, "learning_rate": 8.151891829622688e-07, "loss": 0.9237, "step": 4550 }, { "epoch": 3.0482250502344272, "grad_norm": 3.0274068826784113, "learning_rate": 8.141097305675447e-07, "loss": 0.9177, "step": 4551 }, { "epoch": 3.0488948425987945, "grad_norm": 3.0198087039671253, "learning_rate": 8.130308543203608e-07, "loss": 0.9738, "step": 4552 }, { "epoch": 3.0495646349631613, "grad_norm": 3.103542741747387, "learning_rate": 8.119525545894199e-07, "loss": 0.9009, "step": 4553 }, { "epoch": 3.0502344273275286, "grad_norm": 2.750905930124145, "learning_rate": 8.108748317432291e-07, "loss": 1.0094, "step": 4554 }, { "epoch": 3.0509042196918954, "grad_norm": 2.850333466086498, "learning_rate": 8.097976861500964e-07, "loss": 0.8318, "step": 4555 }, { "epoch": 3.0515740120562627, "grad_norm": 2.9611922294280983, "learning_rate": 8.087211181781307e-07, "loss": 0.9479, "step": 4556 }, { "epoch": 3.0522438044206295, "grad_norm": 3.227338283556582, "learning_rate": 8.076451281952477e-07, "loss": 1.0106, "step": 4557 }, { "epoch": 3.0529135967849967, "grad_norm": 2.9958715805203098, "learning_rate": 8.06569716569163e-07, "loss": 0.9397, "step": 4558 }, { "epoch": 3.0535833891493636, "grad_norm": 2.8592273613941015, "learning_rate": 8.054948836673959e-07, "loss": 0.8623, "step": 4559 }, { "epoch": 3.054253181513731, "grad_norm": 2.8996807627660504, "learning_rate": 8.044206298572643e-07, "loss": 0.9952, "step": 4560 }, { "epoch": 3.0549229738780976, "grad_norm": 2.912866597650047, "learning_rate": 8.033469555058935e-07, "loss": 0.8818, "step": 4561 }, { "epoch": 3.055592766242465, "grad_norm": 2.6414048723170622, "learning_rate": 8.022738609802075e-07, "loss": 0.7524, "step": 4562 }, { "epoch": 3.0562625586068317, "grad_norm": 2.9734034812534165, "learning_rate": 8.012013466469338e-07, "loss": 0.8258, "step": 4563 }, { "epoch": 3.056932350971199, "grad_norm": 2.9224902509945974, "learning_rate": 8.001294128725981e-07, "loss": 1.0101, "step": 4564 }, { "epoch": 3.057602143335566, "grad_norm": 3.306333192722653, "learning_rate": 7.990580600235329e-07, "loss": 0.8926, "step": 4565 }, { "epoch": 3.058271935699933, "grad_norm": 2.9847451189929584, "learning_rate": 7.979872884658685e-07, "loss": 0.9145, "step": 4566 }, { "epoch": 3.0589417280643, "grad_norm": 2.8640937400616, "learning_rate": 7.969170985655383e-07, "loss": 0.7869, "step": 4567 }, { "epoch": 3.059611520428667, "grad_norm": 3.2249495743512155, "learning_rate": 7.958474906882744e-07, "loss": 0.8671, "step": 4568 }, { "epoch": 3.060281312793034, "grad_norm": 2.8432075323465984, "learning_rate": 7.947784651996138e-07, "loss": 0.9547, "step": 4569 }, { "epoch": 3.060951105157401, "grad_norm": 3.017218112317943, "learning_rate": 7.937100224648917e-07, "loss": 1.0278, "step": 4570 }, { "epoch": 3.0616208975217685, "grad_norm": 3.04562490439349, "learning_rate": 7.926421628492461e-07, "loss": 0.9187, "step": 4571 }, { "epoch": 3.0622906898861353, "grad_norm": 2.552123381710475, "learning_rate": 7.915748867176117e-07, "loss": 0.8666, "step": 4572 }, { "epoch": 3.0629604822505025, "grad_norm": 4.233823263842317, "learning_rate": 7.905081944347292e-07, "loss": 0.7796, "step": 4573 }, { "epoch": 3.0636302746148694, "grad_norm": 2.428036956728512, "learning_rate": 7.894420863651367e-07, "loss": 0.7336, "step": 4574 }, { "epoch": 3.0643000669792366, "grad_norm": 3.1856897470984196, "learning_rate": 7.883765628731732e-07, "loss": 0.8308, "step": 4575 }, { "epoch": 3.0649698593436034, "grad_norm": 2.993163085549194, "learning_rate": 7.873116243229773e-07, "loss": 0.8593, "step": 4576 }, { "epoch": 3.0656396517079707, "grad_norm": 2.9761180981691258, "learning_rate": 7.862472710784888e-07, "loss": 0.9745, "step": 4577 }, { "epoch": 3.0663094440723375, "grad_norm": 3.1367263516692137, "learning_rate": 7.851835035034469e-07, "loss": 1.0143, "step": 4578 }, { "epoch": 3.066979236436705, "grad_norm": 3.1347748291631454, "learning_rate": 7.841203219613907e-07, "loss": 0.749, "step": 4579 }, { "epoch": 3.0676490288010716, "grad_norm": 2.998425649994732, "learning_rate": 7.830577268156592e-07, "loss": 1.0123, "step": 4580 }, { "epoch": 3.068318821165439, "grad_norm": 2.572620396966007, "learning_rate": 7.819957184293906e-07, "loss": 0.816, "step": 4581 }, { "epoch": 3.0689886135298057, "grad_norm": 3.198554561740504, "learning_rate": 7.809342971655229e-07, "loss": 0.9548, "step": 4582 }, { "epoch": 3.069658405894173, "grad_norm": 3.382314506260091, "learning_rate": 7.798734633867933e-07, "loss": 0.905, "step": 4583 }, { "epoch": 3.0703281982585398, "grad_norm": 2.785159199198046, "learning_rate": 7.788132174557386e-07, "loss": 0.7792, "step": 4584 }, { "epoch": 3.070997990622907, "grad_norm": 2.9833869102972876, "learning_rate": 7.777535597346941e-07, "loss": 0.8823, "step": 4585 }, { "epoch": 3.071667782987274, "grad_norm": 3.641090727618178, "learning_rate": 7.766944905857946e-07, "loss": 0.8484, "step": 4586 }, { "epoch": 3.072337575351641, "grad_norm": 2.9635673963457054, "learning_rate": 7.756360103709734e-07, "loss": 0.8328, "step": 4587 }, { "epoch": 3.073007367716008, "grad_norm": 2.8959538459566647, "learning_rate": 7.74578119451963e-07, "loss": 0.852, "step": 4588 }, { "epoch": 3.073677160080375, "grad_norm": 2.81570416962511, "learning_rate": 7.735208181902937e-07, "loss": 0.8633, "step": 4589 }, { "epoch": 3.074346952444742, "grad_norm": 4.8626023401603025, "learning_rate": 7.724641069472952e-07, "loss": 0.7613, "step": 4590 }, { "epoch": 3.0750167448091092, "grad_norm": 2.732437449811468, "learning_rate": 7.714079860840948e-07, "loss": 0.7123, "step": 4591 }, { "epoch": 3.075686537173476, "grad_norm": 2.5761978909941416, "learning_rate": 7.703524559616188e-07, "loss": 0.8536, "step": 4592 }, { "epoch": 3.0763563295378433, "grad_norm": 2.9523204660275315, "learning_rate": 7.692975169405909e-07, "loss": 0.915, "step": 4593 }, { "epoch": 3.07702612190221, "grad_norm": 2.850463043511997, "learning_rate": 7.682431693815332e-07, "loss": 0.9407, "step": 4594 }, { "epoch": 3.0776959142665774, "grad_norm": 2.5848721196560924, "learning_rate": 7.671894136447655e-07, "loss": 0.7747, "step": 4595 }, { "epoch": 3.078365706630944, "grad_norm": 2.8437840891292225, "learning_rate": 7.661362500904055e-07, "loss": 0.8739, "step": 4596 }, { "epoch": 3.0790354989953115, "grad_norm": 3.438066271379696, "learning_rate": 7.650836790783681e-07, "loss": 0.8918, "step": 4597 }, { "epoch": 3.0797052913596783, "grad_norm": 2.946912931732851, "learning_rate": 7.640317009683665e-07, "loss": 0.8313, "step": 4598 }, { "epoch": 3.0803750837240456, "grad_norm": 2.972416438042927, "learning_rate": 7.62980316119911e-07, "loss": 0.8802, "step": 4599 }, { "epoch": 3.081044876088413, "grad_norm": 2.970226700828902, "learning_rate": 7.619295248923081e-07, "loss": 0.8901, "step": 4600 }, { "epoch": 3.0817146684527796, "grad_norm": 2.9563012935915656, "learning_rate": 7.60879327644663e-07, "loss": 1.0039, "step": 4601 }, { "epoch": 3.082384460817147, "grad_norm": 3.0719966205524805, "learning_rate": 7.598297247358769e-07, "loss": 1.0174, "step": 4602 }, { "epoch": 3.0830542531815137, "grad_norm": 3.087358420162908, "learning_rate": 7.587807165246483e-07, "loss": 0.923, "step": 4603 }, { "epoch": 3.083724045545881, "grad_norm": 2.952455398230343, "learning_rate": 7.577323033694725e-07, "loss": 0.8747, "step": 4604 }, { "epoch": 3.084393837910248, "grad_norm": 3.0432411981235776, "learning_rate": 7.566844856286412e-07, "loss": 0.9451, "step": 4605 }, { "epoch": 3.085063630274615, "grad_norm": 2.6594930071126455, "learning_rate": 7.556372636602427e-07, "loss": 0.8442, "step": 4606 }, { "epoch": 3.085733422638982, "grad_norm": 2.9746257929169317, "learning_rate": 7.545906378221615e-07, "loss": 0.9225, "step": 4607 }, { "epoch": 3.086403215003349, "grad_norm": 3.122486267944692, "learning_rate": 7.535446084720791e-07, "loss": 1.0019, "step": 4608 }, { "epoch": 3.087073007367716, "grad_norm": 3.2893998560774773, "learning_rate": 7.524991759674721e-07, "loss": 0.9139, "step": 4609 }, { "epoch": 3.087742799732083, "grad_norm": 2.6570815716456946, "learning_rate": 7.514543406656138e-07, "loss": 0.8354, "step": 4610 }, { "epoch": 3.08841259209645, "grad_norm": 3.8539203132798145, "learning_rate": 7.504101029235736e-07, "loss": 0.8703, "step": 4611 }, { "epoch": 3.0890823844608173, "grad_norm": 4.788065109913197, "learning_rate": 7.493664630982153e-07, "loss": 0.8458, "step": 4612 }, { "epoch": 3.089752176825184, "grad_norm": 3.6030030356765583, "learning_rate": 7.48323421546202e-07, "loss": 1.0729, "step": 4613 }, { "epoch": 3.0904219691895514, "grad_norm": 3.0937335203584277, "learning_rate": 7.472809786239871e-07, "loss": 0.8898, "step": 4614 }, { "epoch": 3.091091761553918, "grad_norm": 2.90801371726935, "learning_rate": 7.462391346878222e-07, "loss": 0.985, "step": 4615 }, { "epoch": 3.0917615539182854, "grad_norm": 2.7323714954523797, "learning_rate": 7.451978900937562e-07, "loss": 0.9106, "step": 4616 }, { "epoch": 3.0924313462826523, "grad_norm": 3.1921568869119015, "learning_rate": 7.441572451976303e-07, "loss": 0.9326, "step": 4617 }, { "epoch": 3.0931011386470195, "grad_norm": 3.287739112302673, "learning_rate": 7.431172003550802e-07, "loss": 0.8603, "step": 4618 }, { "epoch": 3.0937709310113863, "grad_norm": 2.8537783789531175, "learning_rate": 7.420777559215381e-07, "loss": 0.9577, "step": 4619 }, { "epoch": 3.0944407233757536, "grad_norm": 2.8469430554482376, "learning_rate": 7.410389122522318e-07, "loss": 0.7572, "step": 4620 }, { "epoch": 3.0951105157401204, "grad_norm": 3.000772035186842, "learning_rate": 7.400006697021836e-07, "loss": 0.9308, "step": 4621 }, { "epoch": 3.0957803081044877, "grad_norm": 3.1631538951802973, "learning_rate": 7.389630286262073e-07, "loss": 1.0326, "step": 4622 }, { "epoch": 3.0964501004688545, "grad_norm": 2.811680967207355, "learning_rate": 7.379259893789134e-07, "loss": 0.846, "step": 4623 }, { "epoch": 3.0971198928332218, "grad_norm": 2.8778871023368597, "learning_rate": 7.368895523147085e-07, "loss": 0.8448, "step": 4624 }, { "epoch": 3.0977896851975886, "grad_norm": 2.963104527178227, "learning_rate": 7.358537177877917e-07, "loss": 0.912, "step": 4625 }, { "epoch": 3.098459477561956, "grad_norm": 2.89510687247251, "learning_rate": 7.34818486152154e-07, "loss": 0.8744, "step": 4626 }, { "epoch": 3.0991292699263226, "grad_norm": 2.8731798455685316, "learning_rate": 7.337838577615833e-07, "loss": 0.8404, "step": 4627 }, { "epoch": 3.09979906229069, "grad_norm": 3.4949358381057225, "learning_rate": 7.327498329696614e-07, "loss": 0.9745, "step": 4628 }, { "epoch": 3.1004688546550567, "grad_norm": 2.7762378295459533, "learning_rate": 7.31716412129763e-07, "loss": 0.9593, "step": 4629 }, { "epoch": 3.101138647019424, "grad_norm": 2.8430410724889508, "learning_rate": 7.30683595595055e-07, "loss": 0.8583, "step": 4630 }, { "epoch": 3.101808439383791, "grad_norm": 3.0067890903322008, "learning_rate": 7.29651383718499e-07, "loss": 0.9211, "step": 4631 }, { "epoch": 3.102478231748158, "grad_norm": 3.199810103061917, "learning_rate": 7.286197768528516e-07, "loss": 0.9584, "step": 4632 }, { "epoch": 3.1031480241125253, "grad_norm": 3.143403511115364, "learning_rate": 7.275887753506611e-07, "loss": 0.8249, "step": 4633 }, { "epoch": 3.103817816476892, "grad_norm": 2.9983815673794427, "learning_rate": 7.265583795642675e-07, "loss": 1.0095, "step": 4634 }, { "epoch": 3.1044876088412594, "grad_norm": 3.270850133745969, "learning_rate": 7.25528589845805e-07, "loss": 1.0452, "step": 4635 }, { "epoch": 3.105157401205626, "grad_norm": 2.9367150845339003, "learning_rate": 7.244994065472024e-07, "loss": 0.7972, "step": 4636 }, { "epoch": 3.1058271935699935, "grad_norm": 3.551151868025844, "learning_rate": 7.234708300201803e-07, "loss": 0.894, "step": 4637 }, { "epoch": 3.1064969859343603, "grad_norm": 2.841373483511014, "learning_rate": 7.224428606162493e-07, "loss": 0.9574, "step": 4638 }, { "epoch": 3.1071667782987276, "grad_norm": 3.002486540576625, "learning_rate": 7.21415498686715e-07, "loss": 0.8319, "step": 4639 }, { "epoch": 3.1078365706630944, "grad_norm": 2.737264380583684, "learning_rate": 7.20388744582676e-07, "loss": 0.9482, "step": 4640 }, { "epoch": 3.1085063630274616, "grad_norm": 2.7879940843431643, "learning_rate": 7.193625986550229e-07, "loss": 0.7836, "step": 4641 }, { "epoch": 3.1091761553918285, "grad_norm": 3.3621717780223968, "learning_rate": 7.183370612544357e-07, "loss": 0.7507, "step": 4642 }, { "epoch": 3.1098459477561957, "grad_norm": 3.0921773755917794, "learning_rate": 7.173121327313887e-07, "loss": 0.8405, "step": 4643 }, { "epoch": 3.1105157401205625, "grad_norm": 3.13966486849141, "learning_rate": 7.162878134361492e-07, "loss": 0.9406, "step": 4644 }, { "epoch": 3.11118553248493, "grad_norm": 3.2004865617779568, "learning_rate": 7.152641037187755e-07, "loss": 0.9475, "step": 4645 }, { "epoch": 3.1118553248492966, "grad_norm": 3.295194647491034, "learning_rate": 7.142410039291142e-07, "loss": 0.8745, "step": 4646 }, { "epoch": 3.112525117213664, "grad_norm": 2.624220668183387, "learning_rate": 7.132185144168089e-07, "loss": 0.9181, "step": 4647 }, { "epoch": 3.1131949095780307, "grad_norm": 2.9076342648997424, "learning_rate": 7.121966355312915e-07, "loss": 0.8103, "step": 4648 }, { "epoch": 3.113864701942398, "grad_norm": 3.2963387570577845, "learning_rate": 7.111753676217856e-07, "loss": 0.8843, "step": 4649 }, { "epoch": 3.1145344943067648, "grad_norm": 3.038416325669196, "learning_rate": 7.101547110373061e-07, "loss": 1.0087, "step": 4650 }, { "epoch": 3.115204286671132, "grad_norm": 2.7922294664144687, "learning_rate": 7.091346661266594e-07, "loss": 0.7222, "step": 4651 }, { "epoch": 3.115874079035499, "grad_norm": 2.7442946629030924, "learning_rate": 7.081152332384419e-07, "loss": 0.8018, "step": 4652 }, { "epoch": 3.116543871399866, "grad_norm": 2.7591407486919906, "learning_rate": 7.070964127210423e-07, "loss": 0.9343, "step": 4653 }, { "epoch": 3.117213663764233, "grad_norm": 2.8096764026069834, "learning_rate": 7.060782049226389e-07, "loss": 0.8375, "step": 4654 }, { "epoch": 3.1178834561286, "grad_norm": 3.3480932880635663, "learning_rate": 7.050606101912008e-07, "loss": 0.9225, "step": 4655 }, { "epoch": 3.118553248492967, "grad_norm": 2.780059092189622, "learning_rate": 7.040436288744879e-07, "loss": 0.8698, "step": 4656 }, { "epoch": 3.1192230408573343, "grad_norm": 2.9658424977067925, "learning_rate": 7.030272613200498e-07, "loss": 0.9125, "step": 4657 }, { "epoch": 3.119892833221701, "grad_norm": 2.8826597958796825, "learning_rate": 7.020115078752273e-07, "loss": 0.8652, "step": 4658 }, { "epoch": 3.1205626255860683, "grad_norm": 3.294769145568628, "learning_rate": 7.00996368887151e-07, "loss": 1.0706, "step": 4659 }, { "epoch": 3.121232417950435, "grad_norm": 3.490697327915268, "learning_rate": 6.99981844702741e-07, "loss": 0.9895, "step": 4660 }, { "epoch": 3.1219022103148024, "grad_norm": 3.091788621047183, "learning_rate": 6.989679356687074e-07, "loss": 0.8947, "step": 4661 }, { "epoch": 3.1225720026791697, "grad_norm": 2.853150911856135, "learning_rate": 6.979546421315511e-07, "loss": 0.8163, "step": 4662 }, { "epoch": 3.1232417950435365, "grad_norm": 3.340263017452517, "learning_rate": 6.96941964437561e-07, "loss": 0.992, "step": 4663 }, { "epoch": 3.1239115874079038, "grad_norm": 3.0086679790677793, "learning_rate": 6.959299029328173e-07, "loss": 0.9816, "step": 4664 }, { "epoch": 3.1245813797722706, "grad_norm": 2.7220844124163395, "learning_rate": 6.949184579631882e-07, "loss": 0.8228, "step": 4665 }, { "epoch": 3.125251172136638, "grad_norm": 2.844652998921034, "learning_rate": 6.939076298743319e-07, "loss": 0.8491, "step": 4666 }, { "epoch": 3.1259209645010047, "grad_norm": 2.8091782397409335, "learning_rate": 6.928974190116955e-07, "loss": 0.9142, "step": 4667 }, { "epoch": 3.126590756865372, "grad_norm": 2.9578170217130313, "learning_rate": 6.918878257205153e-07, "loss": 0.9177, "step": 4668 }, { "epoch": 3.1272605492297387, "grad_norm": 3.4424787873740197, "learning_rate": 6.908788503458166e-07, "loss": 0.7594, "step": 4669 }, { "epoch": 3.127930341594106, "grad_norm": 2.7562541208453153, "learning_rate": 6.898704932324138e-07, "loss": 0.6547, "step": 4670 }, { "epoch": 3.128600133958473, "grad_norm": 2.9238170529274488, "learning_rate": 6.888627547249091e-07, "loss": 0.9973, "step": 4671 }, { "epoch": 3.12926992632284, "grad_norm": 4.084467087014996, "learning_rate": 6.878556351676943e-07, "loss": 0.8388, "step": 4672 }, { "epoch": 3.129939718687207, "grad_norm": 2.909701572537581, "learning_rate": 6.868491349049489e-07, "loss": 0.9639, "step": 4673 }, { "epoch": 3.130609511051574, "grad_norm": 2.7813452038780833, "learning_rate": 6.858432542806414e-07, "loss": 0.9169, "step": 4674 }, { "epoch": 3.131279303415941, "grad_norm": 3.4243701946667757, "learning_rate": 6.848379936385282e-07, "loss": 0.9379, "step": 4675 }, { "epoch": 3.1319490957803082, "grad_norm": 3.4027073152835703, "learning_rate": 6.838333533221539e-07, "loss": 0.9721, "step": 4676 }, { "epoch": 3.132618888144675, "grad_norm": 3.313655430596225, "learning_rate": 6.828293336748507e-07, "loss": 0.9426, "step": 4677 }, { "epoch": 3.1332886805090423, "grad_norm": 3.1081633059956912, "learning_rate": 6.818259350397396e-07, "loss": 0.9492, "step": 4678 }, { "epoch": 3.133958472873409, "grad_norm": 3.42364134197956, "learning_rate": 6.808231577597285e-07, "loss": 0.9799, "step": 4679 }, { "epoch": 3.1346282652377764, "grad_norm": 3.026623487293973, "learning_rate": 6.798210021775134e-07, "loss": 0.9841, "step": 4680 }, { "epoch": 3.135298057602143, "grad_norm": 3.1755755504241616, "learning_rate": 6.788194686355776e-07, "loss": 0.9096, "step": 4681 }, { "epoch": 3.1359678499665105, "grad_norm": 3.270966860893023, "learning_rate": 6.778185574761919e-07, "loss": 0.7876, "step": 4682 }, { "epoch": 3.1366376423308773, "grad_norm": 2.8959700710338003, "learning_rate": 6.76818269041415e-07, "loss": 0.8709, "step": 4683 }, { "epoch": 3.1373074346952445, "grad_norm": 2.8447046180619657, "learning_rate": 6.758186036730915e-07, "loss": 0.7713, "step": 4684 }, { "epoch": 3.1379772270596114, "grad_norm": 2.8790038254179673, "learning_rate": 6.748195617128545e-07, "loss": 0.8771, "step": 4685 }, { "epoch": 3.1386470194239786, "grad_norm": 3.2486328310465615, "learning_rate": 6.738211435021219e-07, "loss": 0.7302, "step": 4686 }, { "epoch": 3.1393168117883454, "grad_norm": 3.1426196849161006, "learning_rate": 6.728233493821023e-07, "loss": 0.8421, "step": 4687 }, { "epoch": 3.1399866041527127, "grad_norm": 3.6640385445176316, "learning_rate": 6.71826179693787e-07, "loss": 0.8718, "step": 4688 }, { "epoch": 3.1406563965170795, "grad_norm": 3.2848024934892646, "learning_rate": 6.708296347779555e-07, "loss": 1.0051, "step": 4689 }, { "epoch": 3.1413261888814468, "grad_norm": 3.192934094473061, "learning_rate": 6.698337149751735e-07, "loss": 0.8125, "step": 4690 }, { "epoch": 3.141995981245814, "grad_norm": 2.9110660694882977, "learning_rate": 6.688384206257955e-07, "loss": 0.7735, "step": 4691 }, { "epoch": 3.142665773610181, "grad_norm": 2.9765622172386323, "learning_rate": 6.678437520699577e-07, "loss": 0.9165, "step": 4692 }, { "epoch": 3.1433355659745477, "grad_norm": 2.66343204629437, "learning_rate": 6.668497096475859e-07, "loss": 0.8876, "step": 4693 }, { "epoch": 3.144005358338915, "grad_norm": 2.9443111933844333, "learning_rate": 6.658562936983898e-07, "loss": 0.8894, "step": 4694 }, { "epoch": 3.144675150703282, "grad_norm": 3.2043484538672335, "learning_rate": 6.648635045618684e-07, "loss": 0.9447, "step": 4695 }, { "epoch": 3.145344943067649, "grad_norm": 3.001599531491674, "learning_rate": 6.638713425773024e-07, "loss": 0.7541, "step": 4696 }, { "epoch": 3.1460147354320163, "grad_norm": 3.1631184583597736, "learning_rate": 6.628798080837601e-07, "loss": 0.8173, "step": 4697 }, { "epoch": 3.146684527796383, "grad_norm": 2.5303328467546677, "learning_rate": 6.618889014200946e-07, "loss": 0.7843, "step": 4698 }, { "epoch": 3.1473543201607503, "grad_norm": 2.6707499549212432, "learning_rate": 6.608986229249476e-07, "loss": 0.8557, "step": 4699 }, { "epoch": 3.148024112525117, "grad_norm": 3.1259680580802054, "learning_rate": 6.599089729367411e-07, "loss": 0.9002, "step": 4700 }, { "epoch": 3.1486939048894844, "grad_norm": 2.624018324879656, "learning_rate": 6.589199517936853e-07, "loss": 0.8104, "step": 4701 }, { "epoch": 3.1493636972538512, "grad_norm": 2.9742827373839686, "learning_rate": 6.579315598337746e-07, "loss": 0.7182, "step": 4702 }, { "epoch": 3.1500334896182185, "grad_norm": 3.11597081929934, "learning_rate": 6.569437973947909e-07, "loss": 0.9405, "step": 4703 }, { "epoch": 3.1507032819825853, "grad_norm": 2.9590876030219952, "learning_rate": 6.559566648142965e-07, "loss": 0.8696, "step": 4704 }, { "epoch": 3.1513730743469526, "grad_norm": 3.103357309371275, "learning_rate": 6.549701624296406e-07, "loss": 0.7685, "step": 4705 }, { "epoch": 3.1520428667113194, "grad_norm": 2.930287317543442, "learning_rate": 6.539842905779589e-07, "loss": 0.7607, "step": 4706 }, { "epoch": 3.1527126590756867, "grad_norm": 2.9739902755816074, "learning_rate": 6.529990495961702e-07, "loss": 0.9741, "step": 4707 }, { "epoch": 3.1533824514400535, "grad_norm": 2.6867843741620177, "learning_rate": 6.520144398209752e-07, "loss": 0.7819, "step": 4708 }, { "epoch": 3.1540522438044207, "grad_norm": 3.1452488811024804, "learning_rate": 6.510304615888618e-07, "loss": 0.7202, "step": 4709 }, { "epoch": 3.1547220361687875, "grad_norm": 2.882237033859045, "learning_rate": 6.500471152361023e-07, "loss": 0.9504, "step": 4710 }, { "epoch": 3.155391828533155, "grad_norm": 2.9449761872636953, "learning_rate": 6.490644010987526e-07, "loss": 0.9788, "step": 4711 }, { "epoch": 3.1560616208975216, "grad_norm": 3.0485438771207627, "learning_rate": 6.480823195126504e-07, "loss": 1.0225, "step": 4712 }, { "epoch": 3.156731413261889, "grad_norm": 2.5790388536427526, "learning_rate": 6.471008708134191e-07, "loss": 0.84, "step": 4713 }, { "epoch": 3.1574012056262557, "grad_norm": 2.7224624026677273, "learning_rate": 6.461200553364666e-07, "loss": 0.9196, "step": 4714 }, { "epoch": 3.158070997990623, "grad_norm": 2.9859268771635756, "learning_rate": 6.451398734169836e-07, "loss": 0.8749, "step": 4715 }, { "epoch": 3.15874079035499, "grad_norm": 2.6241208549075337, "learning_rate": 6.441603253899431e-07, "loss": 0.5596, "step": 4716 }, { "epoch": 3.159410582719357, "grad_norm": 2.907126505085041, "learning_rate": 6.43181411590102e-07, "loss": 0.8607, "step": 4717 }, { "epoch": 3.160080375083724, "grad_norm": 3.023681749666704, "learning_rate": 6.422031323520022e-07, "loss": 0.8448, "step": 4718 }, { "epoch": 3.160750167448091, "grad_norm": 3.376426501922307, "learning_rate": 6.412254880099672e-07, "loss": 0.9442, "step": 4719 }, { "epoch": 3.161419959812458, "grad_norm": 3.0697126781221953, "learning_rate": 6.402484788981045e-07, "loss": 0.8873, "step": 4720 }, { "epoch": 3.162089752176825, "grad_norm": 3.553505985276658, "learning_rate": 6.392721053503012e-07, "loss": 0.921, "step": 4721 }, { "epoch": 3.162759544541192, "grad_norm": 2.9220140047239536, "learning_rate": 6.382963677002324e-07, "loss": 0.8126, "step": 4722 }, { "epoch": 3.1634293369055593, "grad_norm": 3.8635946313121075, "learning_rate": 6.373212662813525e-07, "loss": 0.9179, "step": 4723 }, { "epoch": 3.1640991292699265, "grad_norm": 3.037033235599798, "learning_rate": 6.363468014268997e-07, "loss": 0.8928, "step": 4724 }, { "epoch": 3.1647689216342934, "grad_norm": 2.739503336956794, "learning_rate": 6.353729734698922e-07, "loss": 0.8963, "step": 4725 }, { "epoch": 3.1654387139986606, "grad_norm": 2.947559860136637, "learning_rate": 6.343997827431347e-07, "loss": 0.8752, "step": 4726 }, { "epoch": 3.1661085063630274, "grad_norm": 2.9992444129205733, "learning_rate": 6.334272295792112e-07, "loss": 0.8931, "step": 4727 }, { "epoch": 3.1667782987273947, "grad_norm": 3.0591177634473383, "learning_rate": 6.324553143104898e-07, "loss": 0.9539, "step": 4728 }, { "epoch": 3.1674480910917615, "grad_norm": 2.9762860598227325, "learning_rate": 6.314840372691164e-07, "loss": 1.0459, "step": 4729 }, { "epoch": 3.1681178834561288, "grad_norm": 2.7565350782852804, "learning_rate": 6.305133987870246e-07, "loss": 1.0095, "step": 4730 }, { "epoch": 3.1687876758204956, "grad_norm": 3.0562063405825546, "learning_rate": 6.295433991959258e-07, "loss": 0.8453, "step": 4731 }, { "epoch": 3.169457468184863, "grad_norm": 2.950197222021579, "learning_rate": 6.285740388273154e-07, "loss": 0.8516, "step": 4732 }, { "epoch": 3.1701272605492297, "grad_norm": 2.6733406718078983, "learning_rate": 6.276053180124667e-07, "loss": 0.655, "step": 4733 }, { "epoch": 3.170797052913597, "grad_norm": 3.269937317120999, "learning_rate": 6.266372370824389e-07, "loss": 0.9314, "step": 4734 }, { "epoch": 3.1714668452779637, "grad_norm": 2.9702626378686965, "learning_rate": 6.256697963680702e-07, "loss": 0.8966, "step": 4735 }, { "epoch": 3.172136637642331, "grad_norm": 3.5419366620415267, "learning_rate": 6.247029961999803e-07, "loss": 0.8993, "step": 4736 }, { "epoch": 3.172806430006698, "grad_norm": 2.9658960450745986, "learning_rate": 6.237368369085697e-07, "loss": 0.6143, "step": 4737 }, { "epoch": 3.173476222371065, "grad_norm": 2.453527910581604, "learning_rate": 6.227713188240206e-07, "loss": 0.5366, "step": 4738 }, { "epoch": 3.174146014735432, "grad_norm": 2.6927107037228057, "learning_rate": 6.218064422762954e-07, "loss": 0.7913, "step": 4739 }, { "epoch": 3.174815807099799, "grad_norm": 2.624020463200563, "learning_rate": 6.208422075951376e-07, "loss": 0.9008, "step": 4740 }, { "epoch": 3.175485599464166, "grad_norm": 2.923937714986152, "learning_rate": 6.198786151100713e-07, "loss": 0.9197, "step": 4741 }, { "epoch": 3.1761553918285332, "grad_norm": 3.557542662253071, "learning_rate": 6.189156651504011e-07, "loss": 0.8627, "step": 4742 }, { "epoch": 3.1768251841929, "grad_norm": 3.6950275700726527, "learning_rate": 6.179533580452118e-07, "loss": 0.9918, "step": 4743 }, { "epoch": 3.1774949765572673, "grad_norm": 2.9063179168784417, "learning_rate": 6.169916941233694e-07, "loss": 0.8177, "step": 4744 }, { "epoch": 3.178164768921634, "grad_norm": 3.0369390225148813, "learning_rate": 6.160306737135188e-07, "loss": 0.8362, "step": 4745 }, { "epoch": 3.1788345612860014, "grad_norm": 2.5834340935651534, "learning_rate": 6.150702971440858e-07, "loss": 0.7336, "step": 4746 }, { "epoch": 3.179504353650368, "grad_norm": 3.2417914910758547, "learning_rate": 6.14110564743276e-07, "loss": 1.0175, "step": 4747 }, { "epoch": 3.1801741460147355, "grad_norm": 2.795148935468972, "learning_rate": 6.131514768390748e-07, "loss": 0.8829, "step": 4748 }, { "epoch": 3.1808439383791023, "grad_norm": 3.232916362487775, "learning_rate": 6.121930337592475e-07, "loss": 0.9227, "step": 4749 }, { "epoch": 3.1815137307434695, "grad_norm": 3.0378648846589535, "learning_rate": 6.112352358313389e-07, "loss": 0.8983, "step": 4750 }, { "epoch": 3.1821835231078364, "grad_norm": 3.11537201304134, "learning_rate": 6.102780833826732e-07, "loss": 0.9403, "step": 4751 }, { "epoch": 3.1828533154722036, "grad_norm": 2.7517866008255454, "learning_rate": 6.093215767403543e-07, "loss": 0.8879, "step": 4752 }, { "epoch": 3.183523107836571, "grad_norm": 2.818339387472865, "learning_rate": 6.083657162312653e-07, "loss": 0.8837, "step": 4753 }, { "epoch": 3.1841929002009377, "grad_norm": 3.2188705350575484, "learning_rate": 6.074105021820681e-07, "loss": 0.8756, "step": 4754 }, { "epoch": 3.1848626925653045, "grad_norm": 3.0186905096231142, "learning_rate": 6.064559349192046e-07, "loss": 0.8571, "step": 4755 }, { "epoch": 3.185532484929672, "grad_norm": 2.851201337770573, "learning_rate": 6.055020147688942e-07, "loss": 0.8619, "step": 4756 }, { "epoch": 3.186202277294039, "grad_norm": 3.146271992202121, "learning_rate": 6.045487420571369e-07, "loss": 0.7673, "step": 4757 }, { "epoch": 3.186872069658406, "grad_norm": 3.3369734432522815, "learning_rate": 6.035961171097099e-07, "loss": 0.9242, "step": 4758 }, { "epoch": 3.187541862022773, "grad_norm": 3.695386424603093, "learning_rate": 6.026441402521702e-07, "loss": 1.0043, "step": 4759 }, { "epoch": 3.18821165438714, "grad_norm": 3.0400865089664237, "learning_rate": 6.016928118098525e-07, "loss": 0.9066, "step": 4760 }, { "epoch": 3.188881446751507, "grad_norm": 3.373096597490935, "learning_rate": 6.007421321078705e-07, "loss": 0.7332, "step": 4761 }, { "epoch": 3.189551239115874, "grad_norm": 3.1188395049753543, "learning_rate": 5.997921014711156e-07, "loss": 0.9049, "step": 4762 }, { "epoch": 3.1902210314802413, "grad_norm": 3.0809686376024534, "learning_rate": 5.988427202242575e-07, "loss": 0.9345, "step": 4763 }, { "epoch": 3.190890823844608, "grad_norm": 2.797598639685531, "learning_rate": 5.978939886917442e-07, "loss": 0.8906, "step": 4764 }, { "epoch": 3.1915606162089754, "grad_norm": 3.0285817410197375, "learning_rate": 5.969459071978032e-07, "loss": 0.8594, "step": 4765 }, { "epoch": 3.192230408573342, "grad_norm": 2.595910506940862, "learning_rate": 5.959984760664361e-07, "loss": 0.624, "step": 4766 }, { "epoch": 3.1929002009377094, "grad_norm": 2.7634440231911475, "learning_rate": 5.950516956214255e-07, "loss": 0.7669, "step": 4767 }, { "epoch": 3.1935699933020762, "grad_norm": 3.19972308008324, "learning_rate": 5.941055661863296e-07, "loss": 0.8286, "step": 4768 }, { "epoch": 3.1942397856664435, "grad_norm": 3.2108615699105036, "learning_rate": 5.931600880844874e-07, "loss": 0.8347, "step": 4769 }, { "epoch": 3.1949095780308103, "grad_norm": 3.1099793314910813, "learning_rate": 5.922152616390103e-07, "loss": 0.8117, "step": 4770 }, { "epoch": 3.1955793703951776, "grad_norm": 3.01691234746572, "learning_rate": 5.912710871727911e-07, "loss": 0.9083, "step": 4771 }, { "epoch": 3.1962491627595444, "grad_norm": 2.830351236874401, "learning_rate": 5.903275650084975e-07, "loss": 0.9061, "step": 4772 }, { "epoch": 3.1969189551239117, "grad_norm": 3.9512370570560584, "learning_rate": 5.893846954685767e-07, "loss": 1.0014, "step": 4773 }, { "epoch": 3.1975887474882785, "grad_norm": 3.2309141369825802, "learning_rate": 5.884424788752499e-07, "loss": 0.7815, "step": 4774 }, { "epoch": 3.1982585398526457, "grad_norm": 3.185838834360235, "learning_rate": 5.875009155505168e-07, "loss": 0.9072, "step": 4775 }, { "epoch": 3.1989283322170126, "grad_norm": 2.961700643519111, "learning_rate": 5.865600058161533e-07, "loss": 0.848, "step": 4776 }, { "epoch": 3.19959812458138, "grad_norm": 3.0161641588277264, "learning_rate": 5.856197499937141e-07, "loss": 0.7882, "step": 4777 }, { "epoch": 3.2002679169457466, "grad_norm": 2.968988379948704, "learning_rate": 5.846801484045267e-07, "loss": 0.8727, "step": 4778 }, { "epoch": 3.200937709310114, "grad_norm": 3.057382992626177, "learning_rate": 5.837412013696972e-07, "loss": 0.9758, "step": 4779 }, { "epoch": 3.2016075016744807, "grad_norm": 3.0837886619485837, "learning_rate": 5.828029092101073e-07, "loss": 0.9852, "step": 4780 }, { "epoch": 3.202277294038848, "grad_norm": 3.0812873280457223, "learning_rate": 5.818652722464175e-07, "loss": 0.8156, "step": 4781 }, { "epoch": 3.2029470864032152, "grad_norm": 2.8228490654629295, "learning_rate": 5.809282907990596e-07, "loss": 0.76, "step": 4782 }, { "epoch": 3.203616878767582, "grad_norm": 3.0877469337042607, "learning_rate": 5.799919651882455e-07, "loss": 0.9251, "step": 4783 }, { "epoch": 3.204286671131949, "grad_norm": 3.0946154658480207, "learning_rate": 5.790562957339599e-07, "loss": 0.8327, "step": 4784 }, { "epoch": 3.204956463496316, "grad_norm": 2.860051394094402, "learning_rate": 5.781212827559673e-07, "loss": 0.8497, "step": 4785 }, { "epoch": 3.2056262558606834, "grad_norm": 2.9716874290820123, "learning_rate": 5.771869265738034e-07, "loss": 0.8143, "step": 4786 }, { "epoch": 3.20629604822505, "grad_norm": 2.934813706572177, "learning_rate": 5.76253227506782e-07, "loss": 0.9194, "step": 4787 }, { "epoch": 3.2069658405894175, "grad_norm": 3.582817139213343, "learning_rate": 5.75320185873991e-07, "loss": 0.7291, "step": 4788 }, { "epoch": 3.2076356329537843, "grad_norm": 3.081788747519989, "learning_rate": 5.743878019942964e-07, "loss": 0.8935, "step": 4789 }, { "epoch": 3.2083054253181515, "grad_norm": 3.3872947609905997, "learning_rate": 5.734560761863351e-07, "loss": 1.0422, "step": 4790 }, { "epoch": 3.2089752176825184, "grad_norm": 2.7795440288927558, "learning_rate": 5.725250087685224e-07, "loss": 0.8362, "step": 4791 }, { "epoch": 3.2096450100468856, "grad_norm": 3.290004183506074, "learning_rate": 5.715946000590469e-07, "loss": 0.819, "step": 4792 }, { "epoch": 3.2103148024112524, "grad_norm": 2.97221644625708, "learning_rate": 5.706648503758738e-07, "loss": 0.9982, "step": 4793 }, { "epoch": 3.2109845947756197, "grad_norm": 3.373691094518928, "learning_rate": 5.697357600367423e-07, "loss": 1.0277, "step": 4794 }, { "epoch": 3.2116543871399865, "grad_norm": 3.2943875226500294, "learning_rate": 5.68807329359164e-07, "loss": 0.7694, "step": 4795 }, { "epoch": 3.212324179504354, "grad_norm": 2.772340880780989, "learning_rate": 5.678795586604291e-07, "loss": 1.0168, "step": 4796 }, { "epoch": 3.2129939718687206, "grad_norm": 2.7745769282082247, "learning_rate": 5.669524482575995e-07, "loss": 0.7669, "step": 4797 }, { "epoch": 3.213663764233088, "grad_norm": 3.289876774338865, "learning_rate": 5.660259984675126e-07, "loss": 0.9448, "step": 4798 }, { "epoch": 3.2143335565974547, "grad_norm": 2.882322448693705, "learning_rate": 5.651002096067781e-07, "loss": 1.0161, "step": 4799 }, { "epoch": 3.215003348961822, "grad_norm": 2.92092197755326, "learning_rate": 5.641750819917829e-07, "loss": 0.9004, "step": 4800 }, { "epoch": 3.2156731413261888, "grad_norm": 3.4752800898660094, "learning_rate": 5.632506159386861e-07, "loss": 0.8884, "step": 4801 }, { "epoch": 3.216342933690556, "grad_norm": 2.7270211115318315, "learning_rate": 5.623268117634214e-07, "loss": 0.7861, "step": 4802 }, { "epoch": 3.217012726054923, "grad_norm": 3.1225253833226057, "learning_rate": 5.614036697816938e-07, "loss": 0.8697, "step": 4803 }, { "epoch": 3.21768251841929, "grad_norm": 2.52755980281476, "learning_rate": 5.604811903089863e-07, "loss": 0.6092, "step": 4804 }, { "epoch": 3.218352310783657, "grad_norm": 3.0855171945641025, "learning_rate": 5.595593736605526e-07, "loss": 0.8944, "step": 4805 }, { "epoch": 3.219022103148024, "grad_norm": 3.1042039066283813, "learning_rate": 5.586382201514212e-07, "loss": 0.8591, "step": 4806 }, { "epoch": 3.219691895512391, "grad_norm": 2.714803108509965, "learning_rate": 5.577177300963912e-07, "loss": 0.9262, "step": 4807 }, { "epoch": 3.2203616878767582, "grad_norm": 2.7824753859368103, "learning_rate": 5.56797903810039e-07, "loss": 0.8764, "step": 4808 }, { "epoch": 3.221031480241125, "grad_norm": 2.8980754117941347, "learning_rate": 5.558787416067118e-07, "loss": 0.9339, "step": 4809 }, { "epoch": 3.2217012726054923, "grad_norm": 2.823041993165757, "learning_rate": 5.549602438005311e-07, "loss": 0.6816, "step": 4810 }, { "epoch": 3.222371064969859, "grad_norm": 3.1392205526130432, "learning_rate": 5.540424107053885e-07, "loss": 0.8706, "step": 4811 }, { "epoch": 3.2230408573342264, "grad_norm": 3.0192010657953525, "learning_rate": 5.53125242634952e-07, "loss": 0.8357, "step": 4812 }, { "epoch": 3.2237106496985932, "grad_norm": 3.324732681311597, "learning_rate": 5.522087399026607e-07, "loss": 1.017, "step": 4813 }, { "epoch": 3.2243804420629605, "grad_norm": 3.2206181328157557, "learning_rate": 5.51292902821727e-07, "loss": 1.032, "step": 4814 }, { "epoch": 3.2250502344273277, "grad_norm": 2.6222929782683044, "learning_rate": 5.503777317051332e-07, "loss": 0.6909, "step": 4815 }, { "epoch": 3.2257200267916946, "grad_norm": 3.3832284488687057, "learning_rate": 5.49463226865638e-07, "loss": 0.907, "step": 4816 }, { "epoch": 3.226389819156062, "grad_norm": 2.8161297025694725, "learning_rate": 5.4854938861577e-07, "loss": 0.9569, "step": 4817 }, { "epoch": 3.2270596115204286, "grad_norm": 3.7397708573454524, "learning_rate": 5.476362172678309e-07, "loss": 0.9528, "step": 4818 }, { "epoch": 3.227729403884796, "grad_norm": 2.8012529140331326, "learning_rate": 5.46723713133892e-07, "loss": 0.9565, "step": 4819 }, { "epoch": 3.2283991962491627, "grad_norm": 3.4185460582892166, "learning_rate": 5.458118765258008e-07, "loss": 0.8651, "step": 4820 }, { "epoch": 3.22906898861353, "grad_norm": 2.906637913525918, "learning_rate": 5.449007077551741e-07, "loss": 0.846, "step": 4821 }, { "epoch": 3.229738780977897, "grad_norm": 5.7305988378965305, "learning_rate": 5.439902071334002e-07, "loss": 0.5942, "step": 4822 }, { "epoch": 3.230408573342264, "grad_norm": 2.938773663478041, "learning_rate": 5.430803749716404e-07, "loss": 0.9653, "step": 4823 }, { "epoch": 3.231078365706631, "grad_norm": 2.933190170718744, "learning_rate": 5.421712115808264e-07, "loss": 0.9022, "step": 4824 }, { "epoch": 3.231748158070998, "grad_norm": 2.8170119327575294, "learning_rate": 5.412627172716623e-07, "loss": 0.7149, "step": 4825 }, { "epoch": 3.232417950435365, "grad_norm": 2.7144418071036895, "learning_rate": 5.403548923546228e-07, "loss": 0.7635, "step": 4826 }, { "epoch": 3.233087742799732, "grad_norm": 2.8189029699653445, "learning_rate": 5.394477371399542e-07, "loss": 0.7456, "step": 4827 }, { "epoch": 3.233757535164099, "grad_norm": 2.8782696275429993, "learning_rate": 5.385412519376737e-07, "loss": 0.7941, "step": 4828 }, { "epoch": 3.2344273275284663, "grad_norm": 3.9051887181123344, "learning_rate": 5.376354370575701e-07, "loss": 0.8264, "step": 4829 }, { "epoch": 3.235097119892833, "grad_norm": 3.694080842355683, "learning_rate": 5.367302928092022e-07, "loss": 0.9605, "step": 4830 }, { "epoch": 3.2357669122572004, "grad_norm": 2.6550969102053736, "learning_rate": 5.358258195019003e-07, "loss": 0.6329, "step": 4831 }, { "epoch": 3.236436704621567, "grad_norm": 2.7338951498746886, "learning_rate": 5.349220174447656e-07, "loss": 0.7392, "step": 4832 }, { "epoch": 3.2371064969859344, "grad_norm": 3.075184725069105, "learning_rate": 5.340188869466689e-07, "loss": 0.9678, "step": 4833 }, { "epoch": 3.2377762893503013, "grad_norm": 3.7299972460903987, "learning_rate": 5.331164283162527e-07, "loss": 0.9055, "step": 4834 }, { "epoch": 3.2384460817146685, "grad_norm": 2.8422300598110994, "learning_rate": 5.322146418619287e-07, "loss": 0.6979, "step": 4835 }, { "epoch": 3.2391158740790353, "grad_norm": 2.792547623649376, "learning_rate": 5.313135278918802e-07, "loss": 0.8016, "step": 4836 }, { "epoch": 3.2397856664434026, "grad_norm": 3.0232876109787803, "learning_rate": 5.304130867140591e-07, "loss": 1.0002, "step": 4837 }, { "epoch": 3.2404554588077694, "grad_norm": 3.5557592188573572, "learning_rate": 5.295133186361892e-07, "loss": 0.7841, "step": 4838 }, { "epoch": 3.2411252511721367, "grad_norm": 2.8918568984374122, "learning_rate": 5.28614223965763e-07, "loss": 0.745, "step": 4839 }, { "epoch": 3.2417950435365035, "grad_norm": 3.012206946443956, "learning_rate": 5.277158030100429e-07, "loss": 0.9338, "step": 4840 }, { "epoch": 3.2424648359008708, "grad_norm": 2.9608837937892454, "learning_rate": 5.268180560760614e-07, "loss": 0.8356, "step": 4841 }, { "epoch": 3.2431346282652376, "grad_norm": 3.033480982927891, "learning_rate": 5.259209834706211e-07, "loss": 0.9829, "step": 4842 }, { "epoch": 3.243804420629605, "grad_norm": 2.894638398067027, "learning_rate": 5.250245855002931e-07, "loss": 0.8817, "step": 4843 }, { "epoch": 3.244474212993972, "grad_norm": 3.367011294495049, "learning_rate": 5.241288624714186e-07, "loss": 0.841, "step": 4844 }, { "epoch": 3.245144005358339, "grad_norm": 3.4879366411304416, "learning_rate": 5.232338146901083e-07, "loss": 0.9587, "step": 4845 }, { "epoch": 3.2458137977227057, "grad_norm": 3.8960808747133364, "learning_rate": 5.223394424622419e-07, "loss": 0.7009, "step": 4846 }, { "epoch": 3.246483590087073, "grad_norm": 3.100527858480546, "learning_rate": 5.21445746093468e-07, "loss": 0.9968, "step": 4847 }, { "epoch": 3.2471533824514403, "grad_norm": 3.058187633799563, "learning_rate": 5.205527258892043e-07, "loss": 0.9559, "step": 4848 }, { "epoch": 3.247823174815807, "grad_norm": 2.766468415521803, "learning_rate": 5.196603821546381e-07, "loss": 0.792, "step": 4849 }, { "epoch": 3.2484929671801743, "grad_norm": 3.406154941725055, "learning_rate": 5.187687151947243e-07, "loss": 0.7914, "step": 4850 }, { "epoch": 3.249162759544541, "grad_norm": 3.255900726348161, "learning_rate": 5.178777253141878e-07, "loss": 0.8866, "step": 4851 }, { "epoch": 3.2498325519089084, "grad_norm": 2.548751358421081, "learning_rate": 5.169874128175212e-07, "loss": 0.7782, "step": 4852 }, { "epoch": 3.2505023442732752, "grad_norm": 2.8761834040289673, "learning_rate": 5.160977780089857e-07, "loss": 1.0088, "step": 4853 }, { "epoch": 3.2511721366376425, "grad_norm": 3.3368381441237824, "learning_rate": 5.152088211926109e-07, "loss": 0.9067, "step": 4854 }, { "epoch": 3.2518419290020093, "grad_norm": 3.011603652783681, "learning_rate": 5.14320542672197e-07, "loss": 0.9674, "step": 4855 }, { "epoch": 3.2525117213663766, "grad_norm": 2.897522942351575, "learning_rate": 5.134329427513076e-07, "loss": 0.9419, "step": 4856 }, { "epoch": 3.2531815137307434, "grad_norm": 2.7896870300184817, "learning_rate": 5.125460217332784e-07, "loss": 0.7548, "step": 4857 }, { "epoch": 3.2538513060951106, "grad_norm": 3.233588745991458, "learning_rate": 5.116597799212108e-07, "loss": 0.8727, "step": 4858 }, { "epoch": 3.2545210984594775, "grad_norm": 3.0155835622162517, "learning_rate": 5.107742176179776e-07, "loss": 0.9383, "step": 4859 }, { "epoch": 3.2551908908238447, "grad_norm": 3.3724445763204294, "learning_rate": 5.098893351262141e-07, "loss": 0.8831, "step": 4860 }, { "epoch": 3.2558606831882115, "grad_norm": 3.0394436225864903, "learning_rate": 5.090051327483269e-07, "loss": 0.8697, "step": 4861 }, { "epoch": 3.256530475552579, "grad_norm": 3.1893534859096686, "learning_rate": 5.081216107864892e-07, "loss": 0.9079, "step": 4862 }, { "epoch": 3.2572002679169456, "grad_norm": 3.1226980281311967, "learning_rate": 5.072387695426423e-07, "loss": 0.7159, "step": 4863 }, { "epoch": 3.257870060281313, "grad_norm": 3.5046619487901647, "learning_rate": 5.063566093184949e-07, "loss": 0.987, "step": 4864 }, { "epoch": 3.2585398526456797, "grad_norm": 2.6880539122869394, "learning_rate": 5.054751304155209e-07, "loss": 0.6945, "step": 4865 }, { "epoch": 3.259209645010047, "grad_norm": 3.0519326190540004, "learning_rate": 5.045943331349626e-07, "loss": 0.8352, "step": 4866 }, { "epoch": 3.2598794373744138, "grad_norm": 2.9557286588323506, "learning_rate": 5.037142177778315e-07, "loss": 0.9109, "step": 4867 }, { "epoch": 3.260549229738781, "grad_norm": 3.108939631676174, "learning_rate": 5.028347846449036e-07, "loss": 0.8263, "step": 4868 }, { "epoch": 3.261219022103148, "grad_norm": 2.8003929937748695, "learning_rate": 5.019560340367213e-07, "loss": 0.8864, "step": 4869 }, { "epoch": 3.261888814467515, "grad_norm": 3.0187836125366583, "learning_rate": 5.010779662535947e-07, "loss": 0.8134, "step": 4870 }, { "epoch": 3.262558606831882, "grad_norm": 2.8753506017105708, "learning_rate": 5.002005815956021e-07, "loss": 0.944, "step": 4871 }, { "epoch": 3.263228399196249, "grad_norm": 2.83330027156972, "learning_rate": 4.993238803625869e-07, "loss": 0.7131, "step": 4872 }, { "epoch": 3.2638981915606164, "grad_norm": 3.0059389638095877, "learning_rate": 4.984478628541573e-07, "loss": 0.849, "step": 4873 }, { "epoch": 3.2645679839249833, "grad_norm": 3.339787093716443, "learning_rate": 4.975725293696892e-07, "loss": 0.8292, "step": 4874 }, { "epoch": 3.26523777628935, "grad_norm": 3.1748220186909015, "learning_rate": 4.966978802083272e-07, "loss": 0.944, "step": 4875 }, { "epoch": 3.2659075686537173, "grad_norm": 3.1765414902779403, "learning_rate": 4.95823915668979e-07, "loss": 0.8648, "step": 4876 }, { "epoch": 3.2665773610180846, "grad_norm": 2.99709887765, "learning_rate": 4.949506360503184e-07, "loss": 0.9073, "step": 4877 }, { "epoch": 3.2672471533824514, "grad_norm": 3.0772271234189446, "learning_rate": 4.940780416507854e-07, "loss": 0.8472, "step": 4878 }, { "epoch": 3.2679169457468187, "grad_norm": 2.766929860537147, "learning_rate": 4.932061327685881e-07, "loss": 0.7796, "step": 4879 }, { "epoch": 3.2685867381111855, "grad_norm": 3.033269515694555, "learning_rate": 4.92334909701698e-07, "loss": 0.9392, "step": 4880 }, { "epoch": 3.2692565304755528, "grad_norm": 2.888687283588893, "learning_rate": 4.914643727478519e-07, "loss": 0.8467, "step": 4881 }, { "epoch": 3.2699263228399196, "grad_norm": 3.3503700691882754, "learning_rate": 4.905945222045527e-07, "loss": 0.855, "step": 4882 }, { "epoch": 3.270596115204287, "grad_norm": 4.924001242651299, "learning_rate": 4.897253583690703e-07, "loss": 0.8383, "step": 4883 }, { "epoch": 3.2712659075686537, "grad_norm": 3.425236163743867, "learning_rate": 4.88856881538439e-07, "loss": 0.8861, "step": 4884 }, { "epoch": 3.271935699933021, "grad_norm": 3.361102872505278, "learning_rate": 4.879890920094557e-07, "loss": 1.0085, "step": 4885 }, { "epoch": 3.2726054922973877, "grad_norm": 3.250182097623805, "learning_rate": 4.871219900786866e-07, "loss": 0.8214, "step": 4886 }, { "epoch": 3.273275284661755, "grad_norm": 4.39579371579642, "learning_rate": 4.862555760424606e-07, "loss": 0.9045, "step": 4887 }, { "epoch": 3.273945077026122, "grad_norm": 3.070014929633091, "learning_rate": 4.85389850196872e-07, "loss": 0.8609, "step": 4888 }, { "epoch": 3.274614869390489, "grad_norm": 2.6597317639554174, "learning_rate": 4.845248128377786e-07, "loss": 0.8057, "step": 4889 }, { "epoch": 3.275284661754856, "grad_norm": 2.9905142395210182, "learning_rate": 4.836604642608053e-07, "loss": 0.8946, "step": 4890 }, { "epoch": 3.275954454119223, "grad_norm": 2.566592313787024, "learning_rate": 4.827968047613404e-07, "loss": 0.6558, "step": 4891 }, { "epoch": 3.27662424648359, "grad_norm": 3.9693743998586513, "learning_rate": 4.819338346345371e-07, "loss": 1.003, "step": 4892 }, { "epoch": 3.2772940388479572, "grad_norm": 2.9557805965569455, "learning_rate": 4.810715541753105e-07, "loss": 0.9457, "step": 4893 }, { "epoch": 3.277963831212324, "grad_norm": 3.0194926478397224, "learning_rate": 4.802099636783444e-07, "loss": 0.8142, "step": 4894 }, { "epoch": 3.2786336235766913, "grad_norm": 3.081242681911577, "learning_rate": 4.793490634380837e-07, "loss": 0.904, "step": 4895 }, { "epoch": 3.279303415941058, "grad_norm": 3.0360753203120394, "learning_rate": 4.784888537487395e-07, "loss": 0.8428, "step": 4896 }, { "epoch": 3.2799732083054254, "grad_norm": 2.893347706293264, "learning_rate": 4.776293349042829e-07, "loss": 0.839, "step": 4897 }, { "epoch": 3.280643000669792, "grad_norm": 2.952327266986493, "learning_rate": 4.767705071984541e-07, "loss": 0.8746, "step": 4898 }, { "epoch": 3.2813127930341595, "grad_norm": 3.656218675019402, "learning_rate": 4.75912370924754e-07, "loss": 0.9503, "step": 4899 }, { "epoch": 3.2819825853985263, "grad_norm": 2.755308504487954, "learning_rate": 4.7505492637644784e-07, "loss": 0.9474, "step": 4900 }, { "epoch": 3.2826523777628935, "grad_norm": 2.8803012967001607, "learning_rate": 4.741981738465645e-07, "loss": 0.7951, "step": 4901 }, { "epoch": 3.283322170127261, "grad_norm": 3.1639984928634983, "learning_rate": 4.733421136278962e-07, "loss": 0.7411, "step": 4902 }, { "epoch": 3.2839919624916276, "grad_norm": 2.883311046894586, "learning_rate": 4.724867460129992e-07, "loss": 0.9555, "step": 4903 }, { "epoch": 3.2846617548559944, "grad_norm": 3.5594408430001607, "learning_rate": 4.7163207129419243e-07, "loss": 0.9374, "step": 4904 }, { "epoch": 3.2853315472203617, "grad_norm": 3.120194136003186, "learning_rate": 4.7077808976355805e-07, "loss": 0.966, "step": 4905 }, { "epoch": 3.286001339584729, "grad_norm": 2.877115095017406, "learning_rate": 4.699248017129415e-07, "loss": 0.7149, "step": 4906 }, { "epoch": 3.2866711319490958, "grad_norm": 3.2060666819226262, "learning_rate": 4.6907220743395157e-07, "loss": 0.9205, "step": 4907 }, { "epoch": 3.2873409243134626, "grad_norm": 2.8089377675289127, "learning_rate": 4.6822030721795894e-07, "loss": 0.8907, "step": 4908 }, { "epoch": 3.28801071667783, "grad_norm": 3.061419551751217, "learning_rate": 4.6736910135609827e-07, "loss": 0.9837, "step": 4909 }, { "epoch": 3.288680509042197, "grad_norm": 3.457755923950622, "learning_rate": 4.665185901392663e-07, "loss": 0.8659, "step": 4910 }, { "epoch": 3.289350301406564, "grad_norm": 2.887919819517572, "learning_rate": 4.656687738581225e-07, "loss": 0.8221, "step": 4911 }, { "epoch": 3.290020093770931, "grad_norm": 3.0135709015048153, "learning_rate": 4.648196528030885e-07, "loss": 0.8494, "step": 4912 }, { "epoch": 3.290689886135298, "grad_norm": 3.0999041703635544, "learning_rate": 4.639712272643493e-07, "loss": 0.9141, "step": 4913 }, { "epoch": 3.2913596784996653, "grad_norm": 3.125971881361235, "learning_rate": 4.6312349753185105e-07, "loss": 0.9392, "step": 4914 }, { "epoch": 3.292029470864032, "grad_norm": 3.099341483209411, "learning_rate": 4.6227646389530283e-07, "loss": 1.0725, "step": 4915 }, { "epoch": 3.2926992632283993, "grad_norm": 2.951109429092413, "learning_rate": 4.6143012664417565e-07, "loss": 0.9785, "step": 4916 }, { "epoch": 3.293369055592766, "grad_norm": 3.68899746881395, "learning_rate": 4.6058448606770253e-07, "loss": 0.9879, "step": 4917 }, { "epoch": 3.2940388479571334, "grad_norm": 2.707670322659073, "learning_rate": 4.597395424548784e-07, "loss": 0.6473, "step": 4918 }, { "epoch": 3.2947086403215002, "grad_norm": 3.2017566349383055, "learning_rate": 4.5889529609445984e-07, "loss": 0.9044, "step": 4919 }, { "epoch": 3.2953784326858675, "grad_norm": 2.476669512934662, "learning_rate": 4.5805174727496573e-07, "loss": 0.7148, "step": 4920 }, { "epoch": 3.2960482250502343, "grad_norm": 3.2433565357065484, "learning_rate": 4.572088962846757e-07, "loss": 1.0671, "step": 4921 }, { "epoch": 3.2967180174146016, "grad_norm": 2.81890165067268, "learning_rate": 4.5636674341163143e-07, "loss": 0.8441, "step": 4922 }, { "epoch": 3.2973878097789684, "grad_norm": 3.021011598478182, "learning_rate": 4.555252889436362e-07, "loss": 0.911, "step": 4923 }, { "epoch": 3.2980576021433357, "grad_norm": 3.264025101513278, "learning_rate": 4.5468453316825396e-07, "loss": 0.8775, "step": 4924 }, { "epoch": 3.2987273945077025, "grad_norm": 3.9896014885972306, "learning_rate": 4.5384447637281057e-07, "loss": 0.9708, "step": 4925 }, { "epoch": 3.2993971868720697, "grad_norm": 3.2825891500152973, "learning_rate": 4.5300511884439293e-07, "loss": 0.8788, "step": 4926 }, { "epoch": 3.3000669792364365, "grad_norm": 3.162884971257322, "learning_rate": 4.521664608698481e-07, "loss": 0.8304, "step": 4927 }, { "epoch": 3.300736771600804, "grad_norm": 3.210569719300143, "learning_rate": 4.5132850273578525e-07, "loss": 0.9819, "step": 4928 }, { "epoch": 3.3014065639651706, "grad_norm": 2.8748388191687893, "learning_rate": 4.5049124472857366e-07, "loss": 0.8092, "step": 4929 }, { "epoch": 3.302076356329538, "grad_norm": 3.1535305610267907, "learning_rate": 4.496546871343435e-07, "loss": 1.0022, "step": 4930 }, { "epoch": 3.3027461486939047, "grad_norm": 3.26420669351942, "learning_rate": 4.488188302389859e-07, "loss": 1.0129, "step": 4931 }, { "epoch": 3.303415941058272, "grad_norm": 2.7113843365847177, "learning_rate": 4.4798367432815184e-07, "loss": 0.7063, "step": 4932 }, { "epoch": 3.304085733422639, "grad_norm": 3.0719896765488226, "learning_rate": 4.4714921968725374e-07, "loss": 0.7556, "step": 4933 }, { "epoch": 3.304755525787006, "grad_norm": 2.7827970406062925, "learning_rate": 4.4631546660146296e-07, "loss": 0.7579, "step": 4934 }, { "epoch": 3.3054253181513733, "grad_norm": 3.132511180186929, "learning_rate": 4.4548241535571245e-07, "loss": 1.0359, "step": 4935 }, { "epoch": 3.30609511051574, "grad_norm": 2.7564933008571852, "learning_rate": 4.44650066234695e-07, "loss": 0.7601, "step": 4936 }, { "epoch": 3.306764902880107, "grad_norm": 4.759360102357134, "learning_rate": 4.4381841952286205e-07, "loss": 0.9126, "step": 4937 }, { "epoch": 3.307434695244474, "grad_norm": 3.5539888660814927, "learning_rate": 4.4298747550442805e-07, "loss": 0.8413, "step": 4938 }, { "epoch": 3.3081044876088415, "grad_norm": 2.813307840297228, "learning_rate": 4.421572344633643e-07, "loss": 0.8625, "step": 4939 }, { "epoch": 3.3087742799732083, "grad_norm": 3.167845597208651, "learning_rate": 4.4132769668340265e-07, "loss": 0.9631, "step": 4940 }, { "epoch": 3.3094440723375755, "grad_norm": 3.103377714579086, "learning_rate": 4.4049886244803496e-07, "loss": 0.9353, "step": 4941 }, { "epoch": 3.3101138647019424, "grad_norm": 3.5301718120720964, "learning_rate": 4.3967073204051414e-07, "loss": 0.7337, "step": 4942 }, { "epoch": 3.3107836570663096, "grad_norm": 2.7073737759412126, "learning_rate": 4.3884330574384957e-07, "loss": 0.7123, "step": 4943 }, { "epoch": 3.3114534494306764, "grad_norm": 3.282007632805858, "learning_rate": 4.3801658384081154e-07, "loss": 0.9704, "step": 4944 }, { "epoch": 3.3121232417950437, "grad_norm": 3.0455804202446815, "learning_rate": 4.3719056661393033e-07, "loss": 0.7261, "step": 4945 }, { "epoch": 3.3127930341594105, "grad_norm": 3.6297565417577142, "learning_rate": 4.363652543454949e-07, "loss": 0.9994, "step": 4946 }, { "epoch": 3.3134628265237778, "grad_norm": 2.778363541090523, "learning_rate": 4.355406473175519e-07, "loss": 0.8263, "step": 4947 }, { "epoch": 3.3141326188881446, "grad_norm": 3.0120124848801884, "learning_rate": 4.347167458119081e-07, "loss": 0.8256, "step": 4948 }, { "epoch": 3.314802411252512, "grad_norm": 3.027762266205858, "learning_rate": 4.3389355011013047e-07, "loss": 0.9447, "step": 4949 }, { "epoch": 3.3154722036168787, "grad_norm": 4.038177152204859, "learning_rate": 4.3307106049354316e-07, "loss": 0.9857, "step": 4950 }, { "epoch": 3.316141995981246, "grad_norm": 3.2125342642497743, "learning_rate": 4.322492772432285e-07, "loss": 0.906, "step": 4951 }, { "epoch": 3.3168117883456127, "grad_norm": 3.376311754156001, "learning_rate": 4.314282006400281e-07, "loss": 0.9872, "step": 4952 }, { "epoch": 3.31748158070998, "grad_norm": 4.989860730347937, "learning_rate": 4.3060783096454313e-07, "loss": 0.6389, "step": 4953 }, { "epoch": 3.318151373074347, "grad_norm": 2.877049689746227, "learning_rate": 4.297881684971328e-07, "loss": 0.8834, "step": 4954 }, { "epoch": 3.318821165438714, "grad_norm": 2.9460013328622368, "learning_rate": 4.289692135179127e-07, "loss": 0.9702, "step": 4955 }, { "epoch": 3.319490957803081, "grad_norm": 7.199088267374285, "learning_rate": 4.281509663067579e-07, "loss": 0.6385, "step": 4956 }, { "epoch": 3.320160750167448, "grad_norm": 3.3447011636813304, "learning_rate": 4.27333427143303e-07, "loss": 0.9117, "step": 4957 }, { "epoch": 3.320830542531815, "grad_norm": 2.851220685828411, "learning_rate": 4.265165963069395e-07, "loss": 0.8828, "step": 4958 }, { "epoch": 3.3215003348961822, "grad_norm": 3.1734511104391423, "learning_rate": 4.2570047407681533e-07, "loss": 0.8157, "step": 4959 }, { "epoch": 3.322170127260549, "grad_norm": 2.8286139330897666, "learning_rate": 4.2488506073183775e-07, "loss": 0.8478, "step": 4960 }, { "epoch": 3.3228399196249163, "grad_norm": 3.1079771126651075, "learning_rate": 4.240703565506726e-07, "loss": 0.8195, "step": 4961 }, { "epoch": 3.323509711989283, "grad_norm": 3.6060470040461783, "learning_rate": 4.232563618117427e-07, "loss": 0.883, "step": 4962 }, { "epoch": 3.3241795043536504, "grad_norm": 2.987795211378409, "learning_rate": 4.2244307679322693e-07, "loss": 0.778, "step": 4963 }, { "epoch": 3.3248492967180177, "grad_norm": 3.393109847242874, "learning_rate": 4.2163050177306237e-07, "loss": 0.6602, "step": 4964 }, { "epoch": 3.3255190890823845, "grad_norm": 3.0281819130445875, "learning_rate": 4.208186370289455e-07, "loss": 1.0114, "step": 4965 }, { "epoch": 3.3261888814467513, "grad_norm": 2.9191779816857855, "learning_rate": 4.2000748283832856e-07, "loss": 0.9235, "step": 4966 }, { "epoch": 3.3268586738111185, "grad_norm": 3.0323790620811217, "learning_rate": 4.1919703947841924e-07, "loss": 0.8045, "step": 4967 }, { "epoch": 3.327528466175486, "grad_norm": 3.3655708629133816, "learning_rate": 4.1838730722618406e-07, "loss": 0.8497, "step": 4968 }, { "epoch": 3.3281982585398526, "grad_norm": 2.9942629101563747, "learning_rate": 4.17578286358348e-07, "loss": 0.8556, "step": 4969 }, { "epoch": 3.3288680509042194, "grad_norm": 3.0444385452909866, "learning_rate": 4.1676997715139027e-07, "loss": 0.7956, "step": 4970 }, { "epoch": 3.3295378432685867, "grad_norm": 3.3885079103280193, "learning_rate": 4.1596237988154846e-07, "loss": 0.9755, "step": 4971 }, { "epoch": 3.330207635632954, "grad_norm": 2.636115118202412, "learning_rate": 4.1515549482481515e-07, "loss": 0.7992, "step": 4972 }, { "epoch": 3.330877427997321, "grad_norm": 2.697992696553062, "learning_rate": 4.143493222569417e-07, "loss": 0.8175, "step": 4973 }, { "epoch": 3.331547220361688, "grad_norm": 2.8322874789676136, "learning_rate": 4.135438624534349e-07, "loss": 0.8595, "step": 4974 }, { "epoch": 3.332217012726055, "grad_norm": 3.4080198973509725, "learning_rate": 4.12739115689558e-07, "loss": 0.7961, "step": 4975 }, { "epoch": 3.332886805090422, "grad_norm": 3.1026219064098277, "learning_rate": 4.119350822403304e-07, "loss": 0.7434, "step": 4976 }, { "epoch": 3.333556597454789, "grad_norm": 3.0565561883091092, "learning_rate": 4.1113176238052806e-07, "loss": 0.8824, "step": 4977 }, { "epoch": 3.334226389819156, "grad_norm": 3.330109608773649, "learning_rate": 4.103291563846828e-07, "loss": 0.8094, "step": 4978 }, { "epoch": 3.334896182183523, "grad_norm": 3.3309191337110544, "learning_rate": 4.0952726452708284e-07, "loss": 0.9064, "step": 4979 }, { "epoch": 3.3355659745478903, "grad_norm": 3.2185286303118024, "learning_rate": 4.087260870817719e-07, "loss": 0.9115, "step": 4980 }, { "epoch": 3.336235766912257, "grad_norm": 3.12407940844274, "learning_rate": 4.079256243225502e-07, "loss": 0.9162, "step": 4981 }, { "epoch": 3.3369055592766244, "grad_norm": 2.9655343210301046, "learning_rate": 4.071258765229727e-07, "loss": 0.828, "step": 4982 }, { "epoch": 3.337575351640991, "grad_norm": 3.064209994543248, "learning_rate": 4.0632684395635086e-07, "loss": 0.9308, "step": 4983 }, { "epoch": 3.3382451440053584, "grad_norm": 2.9984701344540685, "learning_rate": 4.055285268957515e-07, "loss": 0.8039, "step": 4984 }, { "epoch": 3.3389149363697253, "grad_norm": 3.850379949916781, "learning_rate": 4.0473092561399694e-07, "loss": 0.8765, "step": 4985 }, { "epoch": 3.3395847287340925, "grad_norm": 2.951197707333678, "learning_rate": 4.039340403836647e-07, "loss": 0.7919, "step": 4986 }, { "epoch": 3.3402545210984593, "grad_norm": 2.8765858585213997, "learning_rate": 4.031378714770881e-07, "loss": 0.892, "step": 4987 }, { "epoch": 3.3409243134628266, "grad_norm": 2.9816561679175546, "learning_rate": 4.0234241916635497e-07, "loss": 0.7314, "step": 4988 }, { "epoch": 3.3415941058271934, "grad_norm": 2.9156383217734176, "learning_rate": 4.0154768372330844e-07, "loss": 0.8891, "step": 4989 }, { "epoch": 3.3422638981915607, "grad_norm": 3.0101001892950703, "learning_rate": 4.0075366541954706e-07, "loss": 0.8828, "step": 4990 }, { "epoch": 3.3429336905559275, "grad_norm": 3.0730677954544494, "learning_rate": 3.999603645264238e-07, "loss": 0.9348, "step": 4991 }, { "epoch": 3.3436034829202947, "grad_norm": 3.2190127093355163, "learning_rate": 3.9916778131504685e-07, "loss": 0.8853, "step": 4992 }, { "epoch": 3.3442732752846616, "grad_norm": 8.012666417968235, "learning_rate": 3.983759160562786e-07, "loss": 0.91, "step": 4993 }, { "epoch": 3.344943067649029, "grad_norm": 3.3632567558109114, "learning_rate": 3.975847690207371e-07, "loss": 0.7504, "step": 4994 }, { "epoch": 3.3456128600133956, "grad_norm": 3.2060441305726077, "learning_rate": 3.9679434047879354e-07, "loss": 0.8471, "step": 4995 }, { "epoch": 3.346282652377763, "grad_norm": 2.947465201072727, "learning_rate": 3.9600463070057493e-07, "loss": 0.6835, "step": 4996 }, { "epoch": 3.34695244474213, "grad_norm": 3.4158038217130318, "learning_rate": 3.952156399559617e-07, "loss": 1.0233, "step": 4997 }, { "epoch": 3.347622237106497, "grad_norm": 3.9645846716426316, "learning_rate": 3.9442736851458914e-07, "loss": 0.8887, "step": 4998 }, { "epoch": 3.348292029470864, "grad_norm": 3.3014948002553277, "learning_rate": 3.9363981664584604e-07, "loss": 0.9054, "step": 4999 }, { "epoch": 3.348961821835231, "grad_norm": 3.192175318917114, "learning_rate": 3.928529846188761e-07, "loss": 1.0147, "step": 5000 }, { "epoch": 3.3496316141995983, "grad_norm": 2.8711942708474423, "learning_rate": 3.9206687270257616e-07, "loss": 0.8772, "step": 5001 }, { "epoch": 3.350301406563965, "grad_norm": 3.1868909246197807, "learning_rate": 3.912814811655979e-07, "loss": 0.8752, "step": 5002 }, { "epoch": 3.3509711989283324, "grad_norm": 4.05805324459294, "learning_rate": 3.904968102763462e-07, "loss": 0.9234, "step": 5003 }, { "epoch": 3.351640991292699, "grad_norm": 3.8367117287007493, "learning_rate": 3.8971286030297983e-07, "loss": 0.982, "step": 5004 }, { "epoch": 3.3523107836570665, "grad_norm": 2.982894551916207, "learning_rate": 3.8892963151341114e-07, "loss": 0.8756, "step": 5005 }, { "epoch": 3.3529805760214333, "grad_norm": 2.7252426451641467, "learning_rate": 3.881471241753057e-07, "loss": 0.7517, "step": 5006 }, { "epoch": 3.3536503683858006, "grad_norm": 3.3392201169247606, "learning_rate": 3.873653385560827e-07, "loss": 0.8997, "step": 5007 }, { "epoch": 3.3543201607501674, "grad_norm": 2.670677385428091, "learning_rate": 3.8658427492291625e-07, "loss": 0.6402, "step": 5008 }, { "epoch": 3.3549899531145346, "grad_norm": 2.962306515423829, "learning_rate": 3.8580393354273107e-07, "loss": 0.96, "step": 5009 }, { "epoch": 3.3556597454789014, "grad_norm": 3.2221583996021472, "learning_rate": 3.8502431468220634e-07, "loss": 0.7877, "step": 5010 }, { "epoch": 3.3563295378432687, "grad_norm": 3.258975443907682, "learning_rate": 3.8424541860777406e-07, "loss": 0.8709, "step": 5011 }, { "epoch": 3.3569993302076355, "grad_norm": 3.1483926009353334, "learning_rate": 3.834672455856209e-07, "loss": 0.944, "step": 5012 }, { "epoch": 3.357669122572003, "grad_norm": 3.9789391458366157, "learning_rate": 3.826897958816833e-07, "loss": 0.8192, "step": 5013 }, { "epoch": 3.3583389149363696, "grad_norm": 3.2894618311643438, "learning_rate": 3.8191306976165327e-07, "loss": 0.8278, "step": 5014 }, { "epoch": 3.359008707300737, "grad_norm": 3.641181019038385, "learning_rate": 3.811370674909731e-07, "loss": 0.8701, "step": 5015 }, { "epoch": 3.3596784996651037, "grad_norm": 3.3768710734440073, "learning_rate": 3.803617893348413e-07, "loss": 0.7904, "step": 5016 }, { "epoch": 3.360348292029471, "grad_norm": 2.531398447377158, "learning_rate": 3.7958723555820445e-07, "loss": 0.7568, "step": 5017 }, { "epoch": 3.3610180843938378, "grad_norm": 3.900631578661751, "learning_rate": 3.788134064257651e-07, "loss": 1.0062, "step": 5018 }, { "epoch": 3.361687876758205, "grad_norm": 3.3784005314023533, "learning_rate": 3.7804030220197575e-07, "loss": 0.8703, "step": 5019 }, { "epoch": 3.362357669122572, "grad_norm": 3.099271372017791, "learning_rate": 3.77267923151044e-07, "loss": 0.7842, "step": 5020 }, { "epoch": 3.363027461486939, "grad_norm": 3.4301710661409404, "learning_rate": 3.7649626953692617e-07, "loss": 0.8938, "step": 5021 }, { "epoch": 3.363697253851306, "grad_norm": 3.2157373045683415, "learning_rate": 3.757253416233334e-07, "loss": 0.7873, "step": 5022 }, { "epoch": 3.364367046215673, "grad_norm": 3.00133004119402, "learning_rate": 3.7495513967372663e-07, "loss": 0.9273, "step": 5023 }, { "epoch": 3.36503683858004, "grad_norm": 3.301965111519269, "learning_rate": 3.7418566395132195e-07, "loss": 0.8776, "step": 5024 }, { "epoch": 3.3657066309444073, "grad_norm": 3.393722600869833, "learning_rate": 3.734169147190833e-07, "loss": 1.0105, "step": 5025 }, { "epoch": 3.3663764233087745, "grad_norm": 3.193130240654118, "learning_rate": 3.7264889223972895e-07, "loss": 0.9686, "step": 5026 }, { "epoch": 3.3670462156731413, "grad_norm": 2.8525487024908975, "learning_rate": 3.718815967757275e-07, "loss": 0.6708, "step": 5027 }, { "epoch": 3.367716008037508, "grad_norm": 3.1522224907565106, "learning_rate": 3.7111502858930144e-07, "loss": 0.8588, "step": 5028 }, { "epoch": 3.3683858004018754, "grad_norm": 3.067262275930982, "learning_rate": 3.703491879424212e-07, "loss": 0.8058, "step": 5029 }, { "epoch": 3.3690555927662427, "grad_norm": 2.8744458455779003, "learning_rate": 3.6958407509681075e-07, "loss": 0.7482, "step": 5030 }, { "epoch": 3.3697253851306095, "grad_norm": 2.8989225551013753, "learning_rate": 3.688196903139449e-07, "loss": 0.8844, "step": 5031 }, { "epoch": 3.3703951774949767, "grad_norm": 2.8986763901542725, "learning_rate": 3.680560338550507e-07, "loss": 0.7423, "step": 5032 }, { "epoch": 3.3710649698593436, "grad_norm": 2.625259859263285, "learning_rate": 3.6729310598110385e-07, "loss": 0.9283, "step": 5033 }, { "epoch": 3.371734762223711, "grad_norm": 3.035374999083243, "learning_rate": 3.665309069528325e-07, "loss": 0.9162, "step": 5034 }, { "epoch": 3.3724045545880776, "grad_norm": 2.8859638479780108, "learning_rate": 3.657694370307169e-07, "loss": 0.7713, "step": 5035 }, { "epoch": 3.373074346952445, "grad_norm": 3.29599412936168, "learning_rate": 3.650086964749869e-07, "loss": 0.9188, "step": 5036 }, { "epoch": 3.3737441393168117, "grad_norm": 3.2599770956867156, "learning_rate": 3.6424868554562203e-07, "loss": 0.8697, "step": 5037 }, { "epoch": 3.374413931681179, "grad_norm": 3.131284359602515, "learning_rate": 3.6348940450235347e-07, "loss": 0.8143, "step": 5038 }, { "epoch": 3.375083724045546, "grad_norm": 3.299317264239342, "learning_rate": 3.6273085360466394e-07, "loss": 0.769, "step": 5039 }, { "epoch": 3.375753516409913, "grad_norm": 2.8813556758672596, "learning_rate": 3.6197303311178627e-07, "loss": 0.8731, "step": 5040 }, { "epoch": 3.37642330877428, "grad_norm": 2.837865578176508, "learning_rate": 3.612159432827017e-07, "loss": 0.9245, "step": 5041 }, { "epoch": 3.377093101138647, "grad_norm": 3.8472897054378006, "learning_rate": 3.6045958437614326e-07, "loss": 0.8679, "step": 5042 }, { "epoch": 3.377762893503014, "grad_norm": 3.1215762253805472, "learning_rate": 3.5970395665059555e-07, "loss": 0.8876, "step": 5043 }, { "epoch": 3.378432685867381, "grad_norm": 3.4887342074072727, "learning_rate": 3.5894906036429087e-07, "loss": 0.7974, "step": 5044 }, { "epoch": 3.379102478231748, "grad_norm": 2.7746130414594226, "learning_rate": 3.581948957752135e-07, "loss": 0.8216, "step": 5045 }, { "epoch": 3.3797722705961153, "grad_norm": 2.992847776920432, "learning_rate": 3.5744146314109495e-07, "loss": 0.816, "step": 5046 }, { "epoch": 3.380442062960482, "grad_norm": 3.475048095048151, "learning_rate": 3.5668876271942014e-07, "loss": 0.9202, "step": 5047 }, { "epoch": 3.3811118553248494, "grad_norm": 3.384775720658819, "learning_rate": 3.559367947674214e-07, "loss": 0.9977, "step": 5048 }, { "epoch": 3.381781647689216, "grad_norm": 2.9415804831840697, "learning_rate": 3.5518555954208195e-07, "loss": 0.8917, "step": 5049 }, { "epoch": 3.3824514400535834, "grad_norm": 3.5596640170539318, "learning_rate": 3.5443505730013205e-07, "loss": 0.9805, "step": 5050 }, { "epoch": 3.3831212324179503, "grad_norm": 3.3083210112679016, "learning_rate": 3.5368528829805516e-07, "loss": 0.9296, "step": 5051 }, { "epoch": 3.3837910247823175, "grad_norm": 3.0685241258353813, "learning_rate": 3.529362527920821e-07, "loss": 0.9529, "step": 5052 }, { "epoch": 3.3844608171466843, "grad_norm": 3.1264753370571707, "learning_rate": 3.52187951038194e-07, "loss": 1.0817, "step": 5053 }, { "epoch": 3.3851306095110516, "grad_norm": 2.8492824003813113, "learning_rate": 3.514403832921184e-07, "loss": 0.8623, "step": 5054 }, { "epoch": 3.385800401875419, "grad_norm": 2.8428619116872045, "learning_rate": 3.506935498093361e-07, "loss": 0.8624, "step": 5055 }, { "epoch": 3.3864701942397857, "grad_norm": 2.8925487536219427, "learning_rate": 3.4994745084507455e-07, "loss": 0.8052, "step": 5056 }, { "epoch": 3.3871399866041525, "grad_norm": 2.96520720622225, "learning_rate": 3.492020866543111e-07, "loss": 0.8859, "step": 5057 }, { "epoch": 3.3878097789685198, "grad_norm": 3.3808568827777976, "learning_rate": 3.4845745749176983e-07, "loss": 0.8178, "step": 5058 }, { "epoch": 3.388479571332887, "grad_norm": 3.353864574782045, "learning_rate": 3.477135636119269e-07, "loss": 1.0167, "step": 5059 }, { "epoch": 3.389149363697254, "grad_norm": 3.3852268036335436, "learning_rate": 3.4697040526900536e-07, "loss": 0.8909, "step": 5060 }, { "epoch": 3.3898191560616207, "grad_norm": 3.3849765190727443, "learning_rate": 3.462279827169773e-07, "loss": 0.9766, "step": 5061 }, { "epoch": 3.390488948425988, "grad_norm": 2.7910189392542, "learning_rate": 3.4548629620956285e-07, "loss": 0.8007, "step": 5062 }, { "epoch": 3.391158740790355, "grad_norm": 3.2234684044143695, "learning_rate": 3.447453460002309e-07, "loss": 0.8055, "step": 5063 }, { "epoch": 3.391828533154722, "grad_norm": 3.082545092147514, "learning_rate": 3.440051323421992e-07, "loss": 0.8114, "step": 5064 }, { "epoch": 3.3924983255190893, "grad_norm": 2.8305958317118316, "learning_rate": 3.432656554884331e-07, "loss": 0.7929, "step": 5065 }, { "epoch": 3.393168117883456, "grad_norm": 2.8819224781963384, "learning_rate": 3.4252691569164636e-07, "loss": 0.8092, "step": 5066 }, { "epoch": 3.3938379102478233, "grad_norm": 2.648387639326342, "learning_rate": 3.417889132043012e-07, "loss": 0.8485, "step": 5067 }, { "epoch": 3.39450770261219, "grad_norm": 3.7852481919147456, "learning_rate": 3.410516482786075e-07, "loss": 1.0, "step": 5068 }, { "epoch": 3.3951774949765574, "grad_norm": 3.0170119320962345, "learning_rate": 3.4031512116652287e-07, "loss": 0.8421, "step": 5069 }, { "epoch": 3.3958472873409242, "grad_norm": 2.7997278868472555, "learning_rate": 3.395793321197535e-07, "loss": 0.8579, "step": 5070 }, { "epoch": 3.3965170797052915, "grad_norm": 2.8976337758760855, "learning_rate": 3.3884428138975247e-07, "loss": 0.8382, "step": 5071 }, { "epoch": 3.3971868720696583, "grad_norm": 2.9229997072762033, "learning_rate": 3.381099692277212e-07, "loss": 0.8301, "step": 5072 }, { "epoch": 3.3978566644340256, "grad_norm": 3.0596905351813137, "learning_rate": 3.373763958846088e-07, "loss": 0.8917, "step": 5073 }, { "epoch": 3.3985264567983924, "grad_norm": 3.573641568248545, "learning_rate": 3.3664356161111106e-07, "loss": 0.9656, "step": 5074 }, { "epoch": 3.3991962491627596, "grad_norm": 3.1580423544042424, "learning_rate": 3.359114666576718e-07, "loss": 1.0007, "step": 5075 }, { "epoch": 3.3998660415271265, "grad_norm": 3.052705215508724, "learning_rate": 3.3518011127448234e-07, "loss": 0.8648, "step": 5076 }, { "epoch": 3.4005358338914937, "grad_norm": 3.1742532322263033, "learning_rate": 3.3444949571148105e-07, "loss": 0.8375, "step": 5077 }, { "epoch": 3.4012056262558605, "grad_norm": 2.9805870996410397, "learning_rate": 3.337196202183532e-07, "loss": 0.856, "step": 5078 }, { "epoch": 3.401875418620228, "grad_norm": 3.2265100267946467, "learning_rate": 3.3299048504453147e-07, "loss": 1.0233, "step": 5079 }, { "epoch": 3.4025452109845946, "grad_norm": 3.2447022612858234, "learning_rate": 3.3226209043919574e-07, "loss": 0.8611, "step": 5080 }, { "epoch": 3.403215003348962, "grad_norm": 4.759725229233227, "learning_rate": 3.3153443665127183e-07, "loss": 0.7417, "step": 5081 }, { "epoch": 3.4038847957133287, "grad_norm": 3.2055968545860414, "learning_rate": 3.308075239294337e-07, "loss": 1.0273, "step": 5082 }, { "epoch": 3.404554588077696, "grad_norm": 3.2705606028164436, "learning_rate": 3.300813525221011e-07, "loss": 0.8671, "step": 5083 }, { "epoch": 3.4052243804420628, "grad_norm": 3.3695122611636195, "learning_rate": 3.2935592267744066e-07, "loss": 0.8586, "step": 5084 }, { "epoch": 3.40589417280643, "grad_norm": 3.119697675633369, "learning_rate": 3.286312346433659e-07, "loss": 0.9441, "step": 5085 }, { "epoch": 3.406563965170797, "grad_norm": 3.3461087510307297, "learning_rate": 3.279072886675369e-07, "loss": 0.8332, "step": 5086 }, { "epoch": 3.407233757535164, "grad_norm": 3.0214426400842576, "learning_rate": 3.2718408499735926e-07, "loss": 0.6961, "step": 5087 }, { "epoch": 3.4079035498995314, "grad_norm": 3.042731579009548, "learning_rate": 3.264616238799856e-07, "loss": 0.895, "step": 5088 }, { "epoch": 3.408573342263898, "grad_norm": 2.797310780765246, "learning_rate": 3.2573990556231503e-07, "loss": 0.8339, "step": 5089 }, { "epoch": 3.409243134628265, "grad_norm": 3.529552985709941, "learning_rate": 3.250189302909923e-07, "loss": 1.0121, "step": 5090 }, { "epoch": 3.4099129269926323, "grad_norm": 2.9595782158471957, "learning_rate": 3.242986983124083e-07, "loss": 0.7271, "step": 5091 }, { "epoch": 3.4105827193569995, "grad_norm": 3.642647962927719, "learning_rate": 3.2357920987269977e-07, "loss": 0.8584, "step": 5092 }, { "epoch": 3.4112525117213663, "grad_norm": 9.24491946459137, "learning_rate": 3.228604652177494e-07, "loss": 0.7774, "step": 5093 }, { "epoch": 3.4119223040857336, "grad_norm": 3.2280051268041445, "learning_rate": 3.221424645931873e-07, "loss": 0.909, "step": 5094 }, { "epoch": 3.4125920964501004, "grad_norm": 2.8909143906656065, "learning_rate": 3.214252082443861e-07, "loss": 0.87, "step": 5095 }, { "epoch": 3.4132618888144677, "grad_norm": 2.967128149638984, "learning_rate": 3.2070869641646656e-07, "loss": 0.8344, "step": 5096 }, { "epoch": 3.4139316811788345, "grad_norm": 2.7883067608859133, "learning_rate": 3.199929293542939e-07, "loss": 0.7947, "step": 5097 }, { "epoch": 3.4146014735432018, "grad_norm": 3.388268143688145, "learning_rate": 3.192779073024804e-07, "loss": 1.0073, "step": 5098 }, { "epoch": 3.4152712659075686, "grad_norm": 2.7617930497900134, "learning_rate": 3.185636305053813e-07, "loss": 0.9291, "step": 5099 }, { "epoch": 3.415941058271936, "grad_norm": 2.852661622557527, "learning_rate": 3.178500992070985e-07, "loss": 0.7691, "step": 5100 }, { "epoch": 3.4166108506363027, "grad_norm": 3.1182592402408296, "learning_rate": 3.171373136514788e-07, "loss": 0.966, "step": 5101 }, { "epoch": 3.41728064300067, "grad_norm": 3.0809443955362137, "learning_rate": 3.164252740821161e-07, "loss": 0.8527, "step": 5102 }, { "epoch": 3.4179504353650367, "grad_norm": 2.9069185430100033, "learning_rate": 3.1571398074234565e-07, "loss": 0.8678, "step": 5103 }, { "epoch": 3.418620227729404, "grad_norm": 2.791877629714803, "learning_rate": 3.150034338752503e-07, "loss": 0.8769, "step": 5104 }, { "epoch": 3.419290020093771, "grad_norm": 2.8918943734884355, "learning_rate": 3.1429363372365655e-07, "loss": 0.7976, "step": 5105 }, { "epoch": 3.419959812458138, "grad_norm": 3.5444070601476287, "learning_rate": 3.13584580530138e-07, "loss": 0.9532, "step": 5106 }, { "epoch": 3.420629604822505, "grad_norm": 4.525908225904779, "learning_rate": 3.128762745370098e-07, "loss": 0.8572, "step": 5107 }, { "epoch": 3.421299397186872, "grad_norm": 3.3021144765292347, "learning_rate": 3.1216871598633346e-07, "loss": 0.6782, "step": 5108 }, { "epoch": 3.421969189551239, "grad_norm": 3.6840142714173205, "learning_rate": 3.114619051199147e-07, "loss": 0.893, "step": 5109 }, { "epoch": 3.4226389819156062, "grad_norm": 3.4632766576862717, "learning_rate": 3.10755842179305e-07, "loss": 1.0984, "step": 5110 }, { "epoch": 3.423308774279973, "grad_norm": 2.7906124693228973, "learning_rate": 3.100505274057977e-07, "loss": 0.8383, "step": 5111 }, { "epoch": 3.4239785666443403, "grad_norm": 2.911924694122619, "learning_rate": 3.0934596104043255e-07, "loss": 0.8769, "step": 5112 }, { "epoch": 3.424648359008707, "grad_norm": 3.3165006618119133, "learning_rate": 3.0864214332399194e-07, "loss": 0.9769, "step": 5113 }, { "epoch": 3.4253181513730744, "grad_norm": 3.0311300686013714, "learning_rate": 3.0793907449700454e-07, "loss": 0.905, "step": 5114 }, { "epoch": 3.425987943737441, "grad_norm": 2.891328262696758, "learning_rate": 3.072367547997418e-07, "loss": 0.8306, "step": 5115 }, { "epoch": 3.4266577361018085, "grad_norm": 2.9134979309811326, "learning_rate": 3.0653518447221837e-07, "loss": 0.8859, "step": 5116 }, { "epoch": 3.4273275284661757, "grad_norm": 2.6801483986374173, "learning_rate": 3.058343637541933e-07, "loss": 0.8112, "step": 5117 }, { "epoch": 3.4279973208305425, "grad_norm": 2.744439344806303, "learning_rate": 3.051342928851711e-07, "loss": 0.8184, "step": 5118 }, { "epoch": 3.4286671131949094, "grad_norm": 3.2326867050160897, "learning_rate": 3.0443497210439847e-07, "loss": 0.7874, "step": 5119 }, { "epoch": 3.4293369055592766, "grad_norm": 3.0105223829751195, "learning_rate": 3.037364016508651e-07, "loss": 0.8432, "step": 5120 }, { "epoch": 3.430006697923644, "grad_norm": 3.1012244303676093, "learning_rate": 3.0303858176330526e-07, "loss": 0.7892, "step": 5121 }, { "epoch": 3.4306764902880107, "grad_norm": 3.177147463129599, "learning_rate": 3.023415126801976e-07, "loss": 0.8338, "step": 5122 }, { "epoch": 3.431346282652378, "grad_norm": 2.890173863648229, "learning_rate": 3.0164519463976334e-07, "loss": 0.8373, "step": 5123 }, { "epoch": 3.4320160750167448, "grad_norm": 2.885214219349418, "learning_rate": 3.0094962787996535e-07, "loss": 0.6453, "step": 5124 }, { "epoch": 3.432685867381112, "grad_norm": 2.980937579756651, "learning_rate": 3.002548126385127e-07, "loss": 0.791, "step": 5125 }, { "epoch": 3.433355659745479, "grad_norm": 4.7840496556893894, "learning_rate": 2.995607491528557e-07, "loss": 0.8327, "step": 5126 }, { "epoch": 3.434025452109846, "grad_norm": 3.2677002996189946, "learning_rate": 2.988674376601891e-07, "loss": 1.024, "step": 5127 }, { "epoch": 3.434695244474213, "grad_norm": 2.9332270138160546, "learning_rate": 2.9817487839744807e-07, "loss": 0.8527, "step": 5128 }, { "epoch": 3.43536503683858, "grad_norm": 3.35015329394799, "learning_rate": 2.974830716013141e-07, "loss": 0.8826, "step": 5129 }, { "epoch": 3.436034829202947, "grad_norm": 3.8965833179677816, "learning_rate": 2.967920175082095e-07, "loss": 0.9274, "step": 5130 }, { "epoch": 3.4367046215673143, "grad_norm": 3.684451286607886, "learning_rate": 2.9610171635430005e-07, "loss": 0.8389, "step": 5131 }, { "epoch": 3.437374413931681, "grad_norm": 2.971473439204424, "learning_rate": 2.954121683754926e-07, "loss": 0.7977, "step": 5132 }, { "epoch": 3.4380442062960483, "grad_norm": 2.781090137226123, "learning_rate": 2.9472337380743954e-07, "loss": 0.8435, "step": 5133 }, { "epoch": 3.438713998660415, "grad_norm": 2.7996796124350194, "learning_rate": 2.9403533288553357e-07, "loss": 0.8051, "step": 5134 }, { "epoch": 3.4393837910247824, "grad_norm": 2.9555193564924096, "learning_rate": 2.933480458449109e-07, "loss": 0.8281, "step": 5135 }, { "epoch": 3.4400535833891492, "grad_norm": 2.8680345896110637, "learning_rate": 2.9266151292044826e-07, "loss": 0.6687, "step": 5136 }, { "epoch": 3.4407233757535165, "grad_norm": 2.8461887261033687, "learning_rate": 2.9197573434676744e-07, "loss": 0.8158, "step": 5137 }, { "epoch": 3.4413931681178833, "grad_norm": 2.7662867004805953, "learning_rate": 2.912907103582305e-07, "loss": 0.7346, "step": 5138 }, { "epoch": 3.4420629604822506, "grad_norm": 3.2677605811992962, "learning_rate": 2.906064411889428e-07, "loss": 0.7597, "step": 5139 }, { "epoch": 3.4427327528466174, "grad_norm": 3.662833215706037, "learning_rate": 2.8992292707274967e-07, "loss": 0.8334, "step": 5140 }, { "epoch": 3.4434025452109847, "grad_norm": 2.8965757880180045, "learning_rate": 2.8924016824324106e-07, "loss": 0.6804, "step": 5141 }, { "epoch": 3.4440723375753515, "grad_norm": 3.494447902738007, "learning_rate": 2.885581649337474e-07, "loss": 0.6957, "step": 5142 }, { "epoch": 3.4447421299397187, "grad_norm": 3.635517544192274, "learning_rate": 2.878769173773416e-07, "loss": 0.8241, "step": 5143 }, { "epoch": 3.4454119223040856, "grad_norm": 3.3063719165829326, "learning_rate": 2.87196425806836e-07, "loss": 0.9171, "step": 5144 }, { "epoch": 3.446081714668453, "grad_norm": 3.1319131514936194, "learning_rate": 2.8651669045478836e-07, "loss": 0.9578, "step": 5145 }, { "epoch": 3.44675150703282, "grad_norm": 2.999986245366305, "learning_rate": 2.8583771155349493e-07, "loss": 0.8952, "step": 5146 }, { "epoch": 3.447421299397187, "grad_norm": 3.149536885271742, "learning_rate": 2.851594893349957e-07, "loss": 0.9866, "step": 5147 }, { "epoch": 3.4480910917615537, "grad_norm": 3.668860324088129, "learning_rate": 2.844820240310689e-07, "loss": 0.8992, "step": 5148 }, { "epoch": 3.448760884125921, "grad_norm": 2.459012158909403, "learning_rate": 2.838053158732379e-07, "loss": 0.726, "step": 5149 }, { "epoch": 3.4494306764902882, "grad_norm": 3.003324959030474, "learning_rate": 2.83129365092765e-07, "loss": 0.8845, "step": 5150 }, { "epoch": 3.450100468854655, "grad_norm": 2.9563827611570095, "learning_rate": 2.824541719206539e-07, "loss": 0.9169, "step": 5151 }, { "epoch": 3.450770261219022, "grad_norm": 3.0992936023884616, "learning_rate": 2.8177973658765015e-07, "loss": 0.9342, "step": 5152 }, { "epoch": 3.451440053583389, "grad_norm": 2.4166306982739565, "learning_rate": 2.8110605932423956e-07, "loss": 0.7758, "step": 5153 }, { "epoch": 3.4521098459477564, "grad_norm": 3.094545378740256, "learning_rate": 2.8043314036064914e-07, "loss": 0.8173, "step": 5154 }, { "epoch": 3.452779638312123, "grad_norm": 2.9184316963423798, "learning_rate": 2.7976097992684665e-07, "loss": 0.8516, "step": 5155 }, { "epoch": 3.4534494306764905, "grad_norm": 3.1680847186487315, "learning_rate": 2.79089578252541e-07, "loss": 0.8995, "step": 5156 }, { "epoch": 3.4541192230408573, "grad_norm": 2.9982510634505304, "learning_rate": 2.7841893556718155e-07, "loss": 0.812, "step": 5157 }, { "epoch": 3.4547890154052245, "grad_norm": 3.0852988024739636, "learning_rate": 2.7774905209995816e-07, "loss": 0.833, "step": 5158 }, { "epoch": 3.4554588077695914, "grad_norm": 2.89371079014531, "learning_rate": 2.7707992807980146e-07, "loss": 0.8131, "step": 5159 }, { "epoch": 3.4561286001339586, "grad_norm": 2.915360715850498, "learning_rate": 2.76411563735382e-07, "loss": 0.8131, "step": 5160 }, { "epoch": 3.4567983924983254, "grad_norm": 3.3561711664755522, "learning_rate": 2.7574395929511165e-07, "loss": 0.985, "step": 5161 }, { "epoch": 3.4574681848626927, "grad_norm": 3.255661084389635, "learning_rate": 2.7507711498714174e-07, "loss": 0.6416, "step": 5162 }, { "epoch": 3.4581379772270595, "grad_norm": 3.674966544968315, "learning_rate": 2.744110310393641e-07, "loss": 0.8837, "step": 5163 }, { "epoch": 3.4588077695914268, "grad_norm": 3.3312128745542653, "learning_rate": 2.737457076794112e-07, "loss": 0.8894, "step": 5164 }, { "epoch": 3.4594775619557936, "grad_norm": 3.0971479156203743, "learning_rate": 2.730811451346546e-07, "loss": 0.8024, "step": 5165 }, { "epoch": 3.460147354320161, "grad_norm": 2.7583044998399635, "learning_rate": 2.7241734363220697e-07, "loss": 0.8567, "step": 5166 }, { "epoch": 3.4608171466845277, "grad_norm": 2.9034560711709987, "learning_rate": 2.717543033989198e-07, "loss": 0.8735, "step": 5167 }, { "epoch": 3.461486939048895, "grad_norm": 3.4901071804556305, "learning_rate": 2.710920246613852e-07, "loss": 0.8953, "step": 5168 }, { "epoch": 3.4621567314132617, "grad_norm": 3.0395761406030046, "learning_rate": 2.704305076459349e-07, "loss": 0.921, "step": 5169 }, { "epoch": 3.462826523777629, "grad_norm": 2.981401762758544, "learning_rate": 2.6976975257864e-07, "loss": 0.9521, "step": 5170 }, { "epoch": 3.463496316141996, "grad_norm": 3.2419547614400757, "learning_rate": 2.691097596853112e-07, "loss": 0.7969, "step": 5171 }, { "epoch": 3.464166108506363, "grad_norm": 2.9068333923052823, "learning_rate": 2.6845052919149934e-07, "loss": 0.7893, "step": 5172 }, { "epoch": 3.46483590087073, "grad_norm": 3.072517737560072, "learning_rate": 2.677920613224941e-07, "loss": 0.7855, "step": 5173 }, { "epoch": 3.465505693235097, "grad_norm": 3.187991318012401, "learning_rate": 2.671343563033249e-07, "loss": 0.8725, "step": 5174 }, { "epoch": 3.466175485599464, "grad_norm": 2.9582586680240324, "learning_rate": 2.6647741435876004e-07, "loss": 0.7845, "step": 5175 }, { "epoch": 3.4668452779638312, "grad_norm": 3.1527738119321884, "learning_rate": 2.6582123571330774e-07, "loss": 0.8793, "step": 5176 }, { "epoch": 3.467515070328198, "grad_norm": 2.5652961970730286, "learning_rate": 2.6516582059121445e-07, "loss": 0.6252, "step": 5177 }, { "epoch": 3.4681848626925653, "grad_norm": 3.539008194321649, "learning_rate": 2.645111692164662e-07, "loss": 0.9118, "step": 5178 }, { "epoch": 3.4688546550569326, "grad_norm": 3.055451955340765, "learning_rate": 2.6385728181278827e-07, "loss": 0.6888, "step": 5179 }, { "epoch": 3.4695244474212994, "grad_norm": 3.419683000394934, "learning_rate": 2.632041586036446e-07, "loss": 0.7862, "step": 5180 }, { "epoch": 3.470194239785666, "grad_norm": 3.252074243952218, "learning_rate": 2.625517998122376e-07, "loss": 0.8934, "step": 5181 }, { "epoch": 3.4708640321500335, "grad_norm": 3.437252203635904, "learning_rate": 2.6190020566150884e-07, "loss": 0.8799, "step": 5182 }, { "epoch": 3.4715338245144007, "grad_norm": 2.751919958238996, "learning_rate": 2.612493763741383e-07, "loss": 0.7563, "step": 5183 }, { "epoch": 3.4722036168787676, "grad_norm": 3.4060552148350918, "learning_rate": 2.6059931217254557e-07, "loss": 0.9028, "step": 5184 }, { "epoch": 3.472873409243135, "grad_norm": 3.2234751105213726, "learning_rate": 2.599500132788879e-07, "loss": 0.9822, "step": 5185 }, { "epoch": 3.4735432016075016, "grad_norm": 3.3342795099678124, "learning_rate": 2.5930147991506003e-07, "loss": 0.9605, "step": 5186 }, { "epoch": 3.474212993971869, "grad_norm": 3.3018612262810807, "learning_rate": 2.5865371230269667e-07, "loss": 0.8929, "step": 5187 }, { "epoch": 3.4748827863362357, "grad_norm": 3.5453976313208586, "learning_rate": 2.580067106631706e-07, "loss": 0.9808, "step": 5188 }, { "epoch": 3.475552578700603, "grad_norm": 3.193889893056904, "learning_rate": 2.573604752175932e-07, "loss": 0.8529, "step": 5189 }, { "epoch": 3.47622237106497, "grad_norm": 2.784737383441281, "learning_rate": 2.5671500618681177e-07, "loss": 0.8276, "step": 5190 }, { "epoch": 3.476892163429337, "grad_norm": 3.7658331240747414, "learning_rate": 2.560703037914136e-07, "loss": 1.0068, "step": 5191 }, { "epoch": 3.477561955793704, "grad_norm": 2.911169198140876, "learning_rate": 2.554263682517244e-07, "loss": 0.9308, "step": 5192 }, { "epoch": 3.478231748158071, "grad_norm": 3.068710532380687, "learning_rate": 2.5478319978780726e-07, "loss": 0.8495, "step": 5193 }, { "epoch": 3.478901540522438, "grad_norm": 3.070567989019932, "learning_rate": 2.541407986194619e-07, "loss": 0.8111, "step": 5194 }, { "epoch": 3.479571332886805, "grad_norm": 4.111883962368841, "learning_rate": 2.534991649662269e-07, "loss": 0.9792, "step": 5195 }, { "epoch": 3.480241125251172, "grad_norm": 3.3354015695177615, "learning_rate": 2.5285829904737934e-07, "loss": 0.9519, "step": 5196 }, { "epoch": 3.4809109176155393, "grad_norm": 3.4137502458875137, "learning_rate": 2.5221820108193297e-07, "loss": 0.8803, "step": 5197 }, { "epoch": 3.481580709979906, "grad_norm": 3.2998568768820276, "learning_rate": 2.5157887128863856e-07, "loss": 0.8392, "step": 5198 }, { "epoch": 3.4822505023442734, "grad_norm": 2.9288439574952894, "learning_rate": 2.509403098859847e-07, "loss": 0.9171, "step": 5199 }, { "epoch": 3.48292029470864, "grad_norm": 2.911715077138946, "learning_rate": 2.5030251709219883e-07, "loss": 0.8567, "step": 5200 }, { "epoch": 3.4835900870730074, "grad_norm": 2.8641910417452374, "learning_rate": 2.496654931252443e-07, "loss": 0.7343, "step": 5201 }, { "epoch": 3.4842598794373743, "grad_norm": 2.779264636593393, "learning_rate": 2.490292382028212e-07, "loss": 0.8275, "step": 5202 }, { "epoch": 3.4849296718017415, "grad_norm": 2.5732405705365924, "learning_rate": 2.483937525423677e-07, "loss": 0.7298, "step": 5203 }, { "epoch": 3.4855994641661083, "grad_norm": 3.004864409799174, "learning_rate": 2.477590363610599e-07, "loss": 0.9898, "step": 5204 }, { "epoch": 3.4862692565304756, "grad_norm": 2.9478847217378874, "learning_rate": 2.471250898758099e-07, "loss": 0.8555, "step": 5205 }, { "epoch": 3.4869390488948424, "grad_norm": 3.2045927724950203, "learning_rate": 2.464919133032659e-07, "loss": 1.0128, "step": 5206 }, { "epoch": 3.4876088412592097, "grad_norm": 2.7332968282954035, "learning_rate": 2.4585950685981405e-07, "loss": 0.8346, "step": 5207 }, { "epoch": 3.488278633623577, "grad_norm": 3.1471922629696825, "learning_rate": 2.4522787076157795e-07, "loss": 0.9051, "step": 5208 }, { "epoch": 3.4889484259879437, "grad_norm": 2.891913263460459, "learning_rate": 2.445970052244173e-07, "loss": 0.8681, "step": 5209 }, { "epoch": 3.4896182183523106, "grad_norm": 3.1113724928593927, "learning_rate": 2.439669104639275e-07, "loss": 0.9378, "step": 5210 }, { "epoch": 3.490288010716678, "grad_norm": 3.026778321547891, "learning_rate": 2.4333758669544156e-07, "loss": 0.9654, "step": 5211 }, { "epoch": 3.490957803081045, "grad_norm": 3.2088699867088173, "learning_rate": 2.4270903413402933e-07, "loss": 0.8404, "step": 5212 }, { "epoch": 3.491627595445412, "grad_norm": 3.5150554385399593, "learning_rate": 2.420812529944969e-07, "loss": 0.8528, "step": 5213 }, { "epoch": 3.4922973878097787, "grad_norm": 3.2406646338163787, "learning_rate": 2.414542434913847e-07, "loss": 0.9308, "step": 5214 }, { "epoch": 3.492967180174146, "grad_norm": 3.0047473500728707, "learning_rate": 2.408280058389731e-07, "loss": 0.8217, "step": 5215 }, { "epoch": 3.4936369725385132, "grad_norm": 3.287177343122226, "learning_rate": 2.4020254025127613e-07, "loss": 0.7202, "step": 5216 }, { "epoch": 3.49430676490288, "grad_norm": 3.1913437902719797, "learning_rate": 2.395778469420451e-07, "loss": 0.8552, "step": 5217 }, { "epoch": 3.4949765572672473, "grad_norm": 3.108722536906218, "learning_rate": 2.389539261247653e-07, "loss": 0.9746, "step": 5218 }, { "epoch": 3.495646349631614, "grad_norm": 3.3762093219152773, "learning_rate": 2.383307780126612e-07, "loss": 0.8905, "step": 5219 }, { "epoch": 3.4963161419959814, "grad_norm": 3.0397667843137572, "learning_rate": 2.3770840281869128e-07, "loss": 0.9215, "step": 5220 }, { "epoch": 3.496985934360348, "grad_norm": 5.065422235724933, "learning_rate": 2.3708680075555045e-07, "loss": 0.9446, "step": 5221 }, { "epoch": 3.4976557267247155, "grad_norm": 2.9026351888928055, "learning_rate": 2.364659720356685e-07, "loss": 0.7501, "step": 5222 }, { "epoch": 3.4983255190890823, "grad_norm": 2.7442756968397095, "learning_rate": 2.358459168712124e-07, "loss": 0.7263, "step": 5223 }, { "epoch": 3.4989953114534496, "grad_norm": 2.6040167992742003, "learning_rate": 2.352266354740837e-07, "loss": 0.7509, "step": 5224 }, { "epoch": 3.4996651038178164, "grad_norm": 3.239667271728208, "learning_rate": 2.3460812805591955e-07, "loss": 0.946, "step": 5225 }, { "epoch": 3.5003348961821836, "grad_norm": 2.9125138696534445, "learning_rate": 2.3399039482809305e-07, "loss": 0.8588, "step": 5226 }, { "epoch": 3.5010046885465504, "grad_norm": 3.075954986967712, "learning_rate": 2.3337343600171286e-07, "loss": 0.9071, "step": 5227 }, { "epoch": 3.5016744809109177, "grad_norm": 3.1895577252902205, "learning_rate": 2.3275725178762225e-07, "loss": 0.9748, "step": 5228 }, { "epoch": 3.5023442732752845, "grad_norm": 3.292278284234185, "learning_rate": 2.3214184239640004e-07, "loss": 0.9954, "step": 5229 }, { "epoch": 3.503014065639652, "grad_norm": 2.9771537361649854, "learning_rate": 2.3152720803836098e-07, "loss": 0.8785, "step": 5230 }, { "epoch": 3.5036838580040186, "grad_norm": 3.0231827314155244, "learning_rate": 2.3091334892355405e-07, "loss": 0.9048, "step": 5231 }, { "epoch": 3.504353650368386, "grad_norm": 3.7213951141997166, "learning_rate": 2.3030026526176336e-07, "loss": 0.9513, "step": 5232 }, { "epoch": 3.5050234427327527, "grad_norm": 3.266993694851331, "learning_rate": 2.2968795726250852e-07, "loss": 1.0369, "step": 5233 }, { "epoch": 3.50569323509712, "grad_norm": 3.2366780071121797, "learning_rate": 2.2907642513504414e-07, "loss": 0.7781, "step": 5234 }, { "epoch": 3.5063630274614868, "grad_norm": 3.4506083997481993, "learning_rate": 2.2846566908835882e-07, "loss": 0.9785, "step": 5235 }, { "epoch": 3.507032819825854, "grad_norm": 2.6285504758653575, "learning_rate": 2.2785568933117647e-07, "loss": 0.7224, "step": 5236 }, { "epoch": 3.5077026121902213, "grad_norm": 2.6433667907571565, "learning_rate": 2.2724648607195615e-07, "loss": 0.9406, "step": 5237 }, { "epoch": 3.508372404554588, "grad_norm": 2.8612325463809136, "learning_rate": 2.266380595188908e-07, "loss": 0.819, "step": 5238 }, { "epoch": 3.509042196918955, "grad_norm": 3.024466196964446, "learning_rate": 2.260304098799082e-07, "loss": 0.8247, "step": 5239 }, { "epoch": 3.509711989283322, "grad_norm": 2.9365103798954544, "learning_rate": 2.2542353736267064e-07, "loss": 0.6866, "step": 5240 }, { "epoch": 3.5103817816476894, "grad_norm": 3.2050194660636206, "learning_rate": 2.2481744217457496e-07, "loss": 0.8941, "step": 5241 }, { "epoch": 3.5110515740120563, "grad_norm": 2.825687779625276, "learning_rate": 2.2421212452275215e-07, "loss": 0.8101, "step": 5242 }, { "epoch": 3.511721366376423, "grad_norm": 2.8953841541596126, "learning_rate": 2.2360758461406757e-07, "loss": 0.918, "step": 5243 }, { "epoch": 3.5123911587407903, "grad_norm": 3.392895972010844, "learning_rate": 2.2300382265512066e-07, "loss": 0.8695, "step": 5244 }, { "epoch": 3.5130609511051576, "grad_norm": 2.754884644795213, "learning_rate": 2.2240083885224556e-07, "loss": 0.8286, "step": 5245 }, { "epoch": 3.5137307434695244, "grad_norm": 2.8373254885725387, "learning_rate": 2.2179863341150936e-07, "loss": 1.0962, "step": 5246 }, { "epoch": 3.5144005358338912, "grad_norm": 3.1027765236049634, "learning_rate": 2.211972065387144e-07, "loss": 0.9273, "step": 5247 }, { "epoch": 3.5150703281982585, "grad_norm": 3.221261985377688, "learning_rate": 2.2059655843939597e-07, "loss": 0.9028, "step": 5248 }, { "epoch": 3.5157401205626257, "grad_norm": 2.753681618897576, "learning_rate": 2.1999668931882407e-07, "loss": 0.8182, "step": 5249 }, { "epoch": 3.5164099129269926, "grad_norm": 2.8539310255010766, "learning_rate": 2.1939759938200134e-07, "loss": 0.6339, "step": 5250 }, { "epoch": 3.51707970529136, "grad_norm": 2.7588715282942005, "learning_rate": 2.187992888336654e-07, "loss": 0.8523, "step": 5251 }, { "epoch": 3.5177494976557266, "grad_norm": 3.0098790755983598, "learning_rate": 2.182017578782869e-07, "loss": 0.7881, "step": 5252 }, { "epoch": 3.518419290020094, "grad_norm": 3.4526072449712717, "learning_rate": 2.1760500672006995e-07, "loss": 0.9124, "step": 5253 }, { "epoch": 3.5190890823844607, "grad_norm": 3.5835656813196093, "learning_rate": 2.170090355629523e-07, "loss": 0.9603, "step": 5254 }, { "epoch": 3.519758874748828, "grad_norm": 3.7278265094855056, "learning_rate": 2.164138446106051e-07, "loss": 0.9712, "step": 5255 }, { "epoch": 3.520428667113195, "grad_norm": 3.2660046797302402, "learning_rate": 2.1581943406643297e-07, "loss": 0.8253, "step": 5256 }, { "epoch": 3.521098459477562, "grad_norm": 5.226583570535258, "learning_rate": 2.1522580413357419e-07, "loss": 0.8485, "step": 5257 }, { "epoch": 3.521768251841929, "grad_norm": 3.124919152992066, "learning_rate": 2.14632955014899e-07, "loss": 0.9242, "step": 5258 }, { "epoch": 3.522438044206296, "grad_norm": 3.0129790558678606, "learning_rate": 2.1404088691301283e-07, "loss": 0.9079, "step": 5259 }, { "epoch": 3.523107836570663, "grad_norm": 3.398948174830733, "learning_rate": 2.1344960003025218e-07, "loss": 0.9419, "step": 5260 }, { "epoch": 3.52377762893503, "grad_norm": 2.4149878949654697, "learning_rate": 2.1285909456868787e-07, "loss": 0.6696, "step": 5261 }, { "epoch": 3.524447421299397, "grad_norm": 3.3117762898140617, "learning_rate": 2.122693707301224e-07, "loss": 0.8779, "step": 5262 }, { "epoch": 3.5251172136637643, "grad_norm": 2.9223399145858036, "learning_rate": 2.116804287160934e-07, "loss": 0.8529, "step": 5263 }, { "epoch": 3.525787006028131, "grad_norm": 2.69373697909782, "learning_rate": 2.1109226872786853e-07, "loss": 0.8423, "step": 5264 }, { "epoch": 3.5264567983924984, "grad_norm": 3.1545958620737187, "learning_rate": 2.105048909664506e-07, "loss": 0.6761, "step": 5265 }, { "epoch": 3.5271265907568656, "grad_norm": 2.9088430986002694, "learning_rate": 2.0991829563257265e-07, "loss": 0.6576, "step": 5266 }, { "epoch": 3.5277963831212324, "grad_norm": 2.838243214016777, "learning_rate": 2.0933248292670406e-07, "loss": 0.8124, "step": 5267 }, { "epoch": 3.5284661754855993, "grad_norm": 2.8654052604140063, "learning_rate": 2.0874745304904216e-07, "loss": 0.8764, "step": 5268 }, { "epoch": 3.5291359678499665, "grad_norm": 2.9097416719327724, "learning_rate": 2.0816320619952008e-07, "loss": 0.8867, "step": 5269 }, { "epoch": 3.529805760214334, "grad_norm": 3.876954133612615, "learning_rate": 2.0757974257780173e-07, "loss": 0.8303, "step": 5270 }, { "epoch": 3.5304755525787006, "grad_norm": 3.416302272163559, "learning_rate": 2.069970623832851e-07, "loss": 0.9446, "step": 5271 }, { "epoch": 3.5311453449430674, "grad_norm": 3.159856248111398, "learning_rate": 2.0641516581509814e-07, "loss": 0.8806, "step": 5272 }, { "epoch": 3.5318151373074347, "grad_norm": 3.4413766691877306, "learning_rate": 2.0583405307210174e-07, "loss": 0.8423, "step": 5273 }, { "epoch": 3.532484929671802, "grad_norm": 2.978226898698106, "learning_rate": 2.052537243528907e-07, "loss": 0.8939, "step": 5274 }, { "epoch": 3.5331547220361688, "grad_norm": 3.9860478480593673, "learning_rate": 2.0467417985578996e-07, "loss": 0.8992, "step": 5275 }, { "epoch": 3.5338245144005356, "grad_norm": 3.3348684954901966, "learning_rate": 2.0409541977885666e-07, "loss": 0.9048, "step": 5276 }, { "epoch": 3.534494306764903, "grad_norm": 2.8232091036256044, "learning_rate": 2.035174443198795e-07, "loss": 0.819, "step": 5277 }, { "epoch": 3.53516409912927, "grad_norm": 2.689658752404452, "learning_rate": 2.0294025367638136e-07, "loss": 0.7857, "step": 5278 }, { "epoch": 3.535833891493637, "grad_norm": 3.142948955502965, "learning_rate": 2.023638480456147e-07, "loss": 0.9049, "step": 5279 }, { "epoch": 3.536503683858004, "grad_norm": 3.659151826988571, "learning_rate": 2.0178822762456362e-07, "loss": 1.0423, "step": 5280 }, { "epoch": 3.537173476222371, "grad_norm": 3.6299017898482973, "learning_rate": 2.0121339260994438e-07, "loss": 0.9732, "step": 5281 }, { "epoch": 3.5378432685867383, "grad_norm": 3.106271058161782, "learning_rate": 2.0063934319820595e-07, "loss": 0.871, "step": 5282 }, { "epoch": 3.538513060951105, "grad_norm": 3.0523014011707534, "learning_rate": 2.0006607958552805e-07, "loss": 1.0032, "step": 5283 }, { "epoch": 3.5391828533154723, "grad_norm": 2.8966759945030205, "learning_rate": 1.9949360196782065e-07, "loss": 0.969, "step": 5284 }, { "epoch": 3.539852645679839, "grad_norm": 2.8307147258388774, "learning_rate": 1.9892191054072612e-07, "loss": 0.8772, "step": 5285 }, { "epoch": 3.5405224380442064, "grad_norm": 2.9796748170123357, "learning_rate": 1.9835100549961878e-07, "loss": 0.914, "step": 5286 }, { "epoch": 3.5411922304085732, "grad_norm": 3.5909800527868847, "learning_rate": 1.977808870396039e-07, "loss": 0.9478, "step": 5287 }, { "epoch": 3.5418620227729405, "grad_norm": 2.8341918589754704, "learning_rate": 1.9721155535551678e-07, "loss": 0.7063, "step": 5288 }, { "epoch": 3.5425318151373073, "grad_norm": 2.685320383477963, "learning_rate": 1.9664301064192453e-07, "loss": 0.7589, "step": 5289 }, { "epoch": 3.5432016075016746, "grad_norm": 2.96294743036483, "learning_rate": 1.960752530931262e-07, "loss": 0.9023, "step": 5290 }, { "epoch": 3.5438713998660414, "grad_norm": 2.966880549739961, "learning_rate": 1.9550828290315155e-07, "loss": 0.7191, "step": 5291 }, { "epoch": 3.5445411922304086, "grad_norm": 2.9455708126489824, "learning_rate": 1.949421002657595e-07, "loss": 0.8016, "step": 5292 }, { "epoch": 3.5452109845947755, "grad_norm": 3.407680794935131, "learning_rate": 1.943767053744411e-07, "loss": 0.7482, "step": 5293 }, { "epoch": 3.5458807769591427, "grad_norm": 3.5580236138221366, "learning_rate": 1.9381209842241927e-07, "loss": 0.8174, "step": 5294 }, { "epoch": 3.54655056932351, "grad_norm": 2.7045232706536937, "learning_rate": 1.9324827960264636e-07, "loss": 0.8072, "step": 5295 }, { "epoch": 3.547220361687877, "grad_norm": 3.41768815843389, "learning_rate": 1.92685249107806e-07, "loss": 0.946, "step": 5296 }, { "epoch": 3.5478901540522436, "grad_norm": 2.840035994956224, "learning_rate": 1.921230071303104e-07, "loss": 0.8174, "step": 5297 }, { "epoch": 3.548559946416611, "grad_norm": 5.158386843330088, "learning_rate": 1.915615538623053e-07, "loss": 0.9455, "step": 5298 }, { "epoch": 3.549229738780978, "grad_norm": 3.294650198684901, "learning_rate": 1.9100088949566552e-07, "loss": 0.8533, "step": 5299 }, { "epoch": 3.549899531145345, "grad_norm": 2.63485685732943, "learning_rate": 1.904410142219959e-07, "loss": 0.6939, "step": 5300 }, { "epoch": 3.5505693235097118, "grad_norm": 3.5718454006826605, "learning_rate": 1.8988192823263197e-07, "loss": 0.7428, "step": 5301 }, { "epoch": 3.551239115874079, "grad_norm": 4.174300095176044, "learning_rate": 1.8932363171864006e-07, "loss": 1.0702, "step": 5302 }, { "epoch": 3.5519089082384463, "grad_norm": 2.9456053194372296, "learning_rate": 1.8876612487081558e-07, "loss": 0.7928, "step": 5303 }, { "epoch": 3.552578700602813, "grad_norm": 2.9570269021974074, "learning_rate": 1.8820940787968507e-07, "loss": 0.9591, "step": 5304 }, { "epoch": 3.55324849296718, "grad_norm": 3.0837420658051187, "learning_rate": 1.8765348093550463e-07, "loss": 0.8825, "step": 5305 }, { "epoch": 3.553918285331547, "grad_norm": 3.346675170284695, "learning_rate": 1.8709834422826035e-07, "loss": 0.8992, "step": 5306 }, { "epoch": 3.5545880776959144, "grad_norm": 2.786405458562901, "learning_rate": 1.8654399794766882e-07, "loss": 0.8093, "step": 5307 }, { "epoch": 3.5552578700602813, "grad_norm": 3.1737099738951042, "learning_rate": 1.859904422831757e-07, "loss": 0.7821, "step": 5308 }, { "epoch": 3.555927662424648, "grad_norm": 6.423177073550757, "learning_rate": 1.8543767742395692e-07, "loss": 0.8921, "step": 5309 }, { "epoch": 3.5565974547890153, "grad_norm": 3.134023906580084, "learning_rate": 1.8488570355891828e-07, "loss": 0.7876, "step": 5310 }, { "epoch": 3.5572672471533826, "grad_norm": 4.094406077023254, "learning_rate": 1.8433452087669474e-07, "loss": 0.8608, "step": 5311 }, { "epoch": 3.5579370395177494, "grad_norm": 3.2084055471449444, "learning_rate": 1.8378412956565145e-07, "loss": 0.9604, "step": 5312 }, { "epoch": 3.5586068318821167, "grad_norm": 3.3458568057937503, "learning_rate": 1.8323452981388267e-07, "loss": 0.7674, "step": 5313 }, { "epoch": 3.5592766242464835, "grad_norm": 2.877442730443462, "learning_rate": 1.8268572180921263e-07, "loss": 0.7667, "step": 5314 }, { "epoch": 3.5599464166108508, "grad_norm": 2.966801351842841, "learning_rate": 1.8213770573919433e-07, "loss": 0.8698, "step": 5315 }, { "epoch": 3.5606162089752176, "grad_norm": 2.884084225723718, "learning_rate": 1.8159048179111072e-07, "loss": 0.903, "step": 5316 }, { "epoch": 3.561286001339585, "grad_norm": 2.928759360163599, "learning_rate": 1.810440501519739e-07, "loss": 0.8681, "step": 5317 }, { "epoch": 3.5619557937039517, "grad_norm": 3.0705063149939797, "learning_rate": 1.8049841100852477e-07, "loss": 1.0057, "step": 5318 }, { "epoch": 3.562625586068319, "grad_norm": 3.455001304879309, "learning_rate": 1.7995356454723416e-07, "loss": 0.9466, "step": 5319 }, { "epoch": 3.5632953784326857, "grad_norm": 3.0440594876800735, "learning_rate": 1.794095109543012e-07, "loss": 0.9194, "step": 5320 }, { "epoch": 3.563965170797053, "grad_norm": 3.105121355822949, "learning_rate": 1.7886625041565465e-07, "loss": 0.8947, "step": 5321 }, { "epoch": 3.56463496316142, "grad_norm": 3.1859044960231877, "learning_rate": 1.783237831169521e-07, "loss": 0.7823, "step": 5322 }, { "epoch": 3.565304755525787, "grad_norm": 3.729054299048746, "learning_rate": 1.777821092435797e-07, "loss": 0.9104, "step": 5323 }, { "epoch": 3.565974547890154, "grad_norm": 3.8744371013440557, "learning_rate": 1.77241228980653e-07, "loss": 0.886, "step": 5324 }, { "epoch": 3.566644340254521, "grad_norm": 2.9660907015481084, "learning_rate": 1.7670114251301607e-07, "loss": 0.8706, "step": 5325 }, { "epoch": 3.567314132618888, "grad_norm": 3.3100207687877536, "learning_rate": 1.7616185002524155e-07, "loss": 0.9575, "step": 5326 }, { "epoch": 3.5679839249832552, "grad_norm": 3.1587084091224145, "learning_rate": 1.7562335170163091e-07, "loss": 0.8909, "step": 5327 }, { "epoch": 3.5686537173476225, "grad_norm": 3.246973034068457, "learning_rate": 1.7508564772621444e-07, "loss": 0.8597, "step": 5328 }, { "epoch": 3.5693235097119893, "grad_norm": 3.077736119060602, "learning_rate": 1.745487382827507e-07, "loss": 1.0482, "step": 5329 }, { "epoch": 3.569993302076356, "grad_norm": 2.809379908118795, "learning_rate": 1.7401262355472653e-07, "loss": 0.6588, "step": 5330 }, { "epoch": 3.5706630944407234, "grad_norm": 3.062933173202512, "learning_rate": 1.734773037253576e-07, "loss": 0.7524, "step": 5331 }, { "epoch": 3.5713328868050906, "grad_norm": 3.218596087631142, "learning_rate": 1.7294277897758728e-07, "loss": 0.9448, "step": 5332 }, { "epoch": 3.5720026791694575, "grad_norm": 3.0430416325871312, "learning_rate": 1.724090494940886e-07, "loss": 0.7815, "step": 5333 }, { "epoch": 3.5726724715338243, "grad_norm": 3.096446210554396, "learning_rate": 1.7187611545726125e-07, "loss": 0.8244, "step": 5334 }, { "epoch": 3.5733422638981915, "grad_norm": 3.2844508651962356, "learning_rate": 1.7134397704923366e-07, "loss": 0.884, "step": 5335 }, { "epoch": 3.574012056262559, "grad_norm": 3.0555857553929147, "learning_rate": 1.708126344518621e-07, "loss": 0.6751, "step": 5336 }, { "epoch": 3.5746818486269256, "grad_norm": 3.1230621836957577, "learning_rate": 1.7028208784673234e-07, "loss": 0.7266, "step": 5337 }, { "epoch": 3.5753516409912924, "grad_norm": 3.145612510907025, "learning_rate": 1.6975233741515607e-07, "loss": 0.9524, "step": 5338 }, { "epoch": 3.5760214333556597, "grad_norm": 2.876961988753148, "learning_rate": 1.6922338333817374e-07, "loss": 0.9561, "step": 5339 }, { "epoch": 3.576691225720027, "grad_norm": 3.298835571315476, "learning_rate": 1.686952257965535e-07, "loss": 0.8945, "step": 5340 }, { "epoch": 3.5773610180843938, "grad_norm": 3.49700614737299, "learning_rate": 1.6816786497079268e-07, "loss": 1.0206, "step": 5341 }, { "epoch": 3.578030810448761, "grad_norm": 3.8192563380047715, "learning_rate": 1.6764130104111404e-07, "loss": 0.9361, "step": 5342 }, { "epoch": 3.578700602813128, "grad_norm": 3.3286187988539044, "learning_rate": 1.6711553418746946e-07, "loss": 0.9028, "step": 5343 }, { "epoch": 3.579370395177495, "grad_norm": 3.062663227058825, "learning_rate": 1.665905645895377e-07, "loss": 0.7863, "step": 5344 }, { "epoch": 3.580040187541862, "grad_norm": 3.086297772617859, "learning_rate": 1.660663924267264e-07, "loss": 0.8049, "step": 5345 }, { "epoch": 3.580709979906229, "grad_norm": 3.549279857560531, "learning_rate": 1.6554301787816863e-07, "loss": 0.9347, "step": 5346 }, { "epoch": 3.581379772270596, "grad_norm": 2.992424529836046, "learning_rate": 1.650204411227263e-07, "loss": 0.7726, "step": 5347 }, { "epoch": 3.5820495646349633, "grad_norm": 3.2855147371501108, "learning_rate": 1.6449866233898797e-07, "loss": 1.0836, "step": 5348 }, { "epoch": 3.58271935699933, "grad_norm": 3.058986996423173, "learning_rate": 1.6397768170527102e-07, "loss": 0.9246, "step": 5349 }, { "epoch": 3.5833891493636973, "grad_norm": 3.3059719315167655, "learning_rate": 1.634574993996177e-07, "loss": 0.9737, "step": 5350 }, { "epoch": 3.584058941728064, "grad_norm": 3.0774350256378837, "learning_rate": 1.6293811559979916e-07, "loss": 0.9649, "step": 5351 }, { "epoch": 3.5847287340924314, "grad_norm": 3.225752218827939, "learning_rate": 1.6241953048331232e-07, "loss": 0.8061, "step": 5352 }, { "epoch": 3.5853985264567982, "grad_norm": 2.8224803161380536, "learning_rate": 1.6190174422738376e-07, "loss": 0.7478, "step": 5353 }, { "epoch": 3.5860683188211655, "grad_norm": 2.5268057614417163, "learning_rate": 1.6138475700896384e-07, "loss": 0.5697, "step": 5354 }, { "epoch": 3.5867381111855323, "grad_norm": 3.3916062038263055, "learning_rate": 1.6086856900473153e-07, "loss": 0.8583, "step": 5355 }, { "epoch": 3.5874079035498996, "grad_norm": 2.9746460988534933, "learning_rate": 1.603531803910918e-07, "loss": 0.8105, "step": 5356 }, { "epoch": 3.588077695914267, "grad_norm": 3.6874477643508694, "learning_rate": 1.5983859134417885e-07, "loss": 0.8177, "step": 5357 }, { "epoch": 3.5887474882786337, "grad_norm": 3.269447095417689, "learning_rate": 1.5932480203984996e-07, "loss": 0.8042, "step": 5358 }, { "epoch": 3.5894172806430005, "grad_norm": 3.081475474756438, "learning_rate": 1.5881181265369195e-07, "loss": 0.8921, "step": 5359 }, { "epoch": 3.5900870730073677, "grad_norm": 2.96099390781878, "learning_rate": 1.582996233610165e-07, "loss": 0.9419, "step": 5360 }, { "epoch": 3.590756865371735, "grad_norm": 3.0497383790914103, "learning_rate": 1.577882343368639e-07, "loss": 0.8739, "step": 5361 }, { "epoch": 3.591426657736102, "grad_norm": 3.143533953105429, "learning_rate": 1.572776457559988e-07, "loss": 0.8257, "step": 5362 }, { "epoch": 3.5920964501004686, "grad_norm": 2.8094249712015027, "learning_rate": 1.567678577929127e-07, "loss": 0.765, "step": 5363 }, { "epoch": 3.592766242464836, "grad_norm": 3.0881039313887615, "learning_rate": 1.5625887062182515e-07, "loss": 0.8746, "step": 5364 }, { "epoch": 3.593436034829203, "grad_norm": 2.7352186843856185, "learning_rate": 1.5575068441668033e-07, "loss": 0.863, "step": 5365 }, { "epoch": 3.59410582719357, "grad_norm": 2.881382467967215, "learning_rate": 1.5524329935114961e-07, "loss": 0.7518, "step": 5366 }, { "epoch": 3.594775619557937, "grad_norm": 3.2221802246085893, "learning_rate": 1.54736715598629e-07, "loss": 0.926, "step": 5367 }, { "epoch": 3.595445411922304, "grad_norm": 3.2186497254083446, "learning_rate": 1.5423093333224308e-07, "loss": 0.8433, "step": 5368 }, { "epoch": 3.5961152042866713, "grad_norm": 3.3662711294149874, "learning_rate": 1.5372595272484082e-07, "loss": 0.9649, "step": 5369 }, { "epoch": 3.596784996651038, "grad_norm": 3.1269621747510263, "learning_rate": 1.532217739489983e-07, "loss": 0.8112, "step": 5370 }, { "epoch": 3.5974547890154054, "grad_norm": 3.0597290045778847, "learning_rate": 1.5271839717701577e-07, "loss": 0.8307, "step": 5371 }, { "epoch": 3.598124581379772, "grad_norm": 2.705589631205448, "learning_rate": 1.5221582258092173e-07, "loss": 0.7183, "step": 5372 }, { "epoch": 3.5987943737441395, "grad_norm": 3.299162662732675, "learning_rate": 1.5171405033246906e-07, "loss": 0.7505, "step": 5373 }, { "epoch": 3.5994641661085063, "grad_norm": 3.2183005181595536, "learning_rate": 1.51213080603137e-07, "loss": 0.7796, "step": 5374 }, { "epoch": 3.6001339584728735, "grad_norm": 3.095720417215186, "learning_rate": 1.5071291356412914e-07, "loss": 0.7563, "step": 5375 }, { "epoch": 3.6008037508372404, "grad_norm": 2.8526992147391335, "learning_rate": 1.5021354938637766e-07, "loss": 0.8261, "step": 5376 }, { "epoch": 3.6014735432016076, "grad_norm": 3.064542770909534, "learning_rate": 1.4971498824053771e-07, "loss": 0.9546, "step": 5377 }, { "epoch": 3.6021433355659744, "grad_norm": 2.8704991346939974, "learning_rate": 1.4921723029699132e-07, "loss": 0.8267, "step": 5378 }, { "epoch": 3.6028131279303417, "grad_norm": 2.94710016852367, "learning_rate": 1.4872027572584492e-07, "loss": 0.8651, "step": 5379 }, { "epoch": 3.6034829202947085, "grad_norm": 3.3054691834460272, "learning_rate": 1.4822412469693208e-07, "loss": 0.914, "step": 5380 }, { "epoch": 3.6041527126590758, "grad_norm": 2.9313551122188777, "learning_rate": 1.4772877737981029e-07, "loss": 0.7814, "step": 5381 }, { "epoch": 3.6048225050234426, "grad_norm": 3.4063121073078713, "learning_rate": 1.472342339437635e-07, "loss": 0.9656, "step": 5382 }, { "epoch": 3.60549229738781, "grad_norm": 3.054574884777332, "learning_rate": 1.467404945577991e-07, "loss": 0.8485, "step": 5383 }, { "epoch": 3.6061620897521767, "grad_norm": 2.973845274143771, "learning_rate": 1.4624755939065177e-07, "loss": 0.8128, "step": 5384 }, { "epoch": 3.606831882116544, "grad_norm": 2.874422101513379, "learning_rate": 1.457554286107804e-07, "loss": 0.8341, "step": 5385 }, { "epoch": 3.607501674480911, "grad_norm": 2.9511690580987784, "learning_rate": 1.4526410238636963e-07, "loss": 0.8226, "step": 5386 }, { "epoch": 3.608171466845278, "grad_norm": 3.2334944894435904, "learning_rate": 1.4477358088532734e-07, "loss": 0.9164, "step": 5387 }, { "epoch": 3.608841259209645, "grad_norm": 3.2169022703216057, "learning_rate": 1.4428386427528863e-07, "loss": 0.8458, "step": 5388 }, { "epoch": 3.609511051574012, "grad_norm": 2.867102065768548, "learning_rate": 1.4379495272361237e-07, "loss": 0.8004, "step": 5389 }, { "epoch": 3.6101808439383793, "grad_norm": 3.0917449947327333, "learning_rate": 1.4330684639738246e-07, "loss": 0.9396, "step": 5390 }, { "epoch": 3.610850636302746, "grad_norm": 3.0881347955179805, "learning_rate": 1.4281954546340764e-07, "loss": 0.8345, "step": 5391 }, { "epoch": 3.611520428667113, "grad_norm": 2.8958378815463957, "learning_rate": 1.423330500882214e-07, "loss": 0.7707, "step": 5392 }, { "epoch": 3.6121902210314802, "grad_norm": 2.88680225443573, "learning_rate": 1.4184736043808184e-07, "loss": 0.7622, "step": 5393 }, { "epoch": 3.6128600133958475, "grad_norm": 3.058624504399893, "learning_rate": 1.4136247667897203e-07, "loss": 0.8625, "step": 5394 }, { "epoch": 3.6135298057602143, "grad_norm": 2.9865055213921945, "learning_rate": 1.4087839897659967e-07, "loss": 0.862, "step": 5395 }, { "epoch": 3.614199598124581, "grad_norm": 2.771027579402424, "learning_rate": 1.4039512749639605e-07, "loss": 0.9209, "step": 5396 }, { "epoch": 3.6148693904889484, "grad_norm": 3.1543677469558378, "learning_rate": 1.3991266240351825e-07, "loss": 0.7756, "step": 5397 }, { "epoch": 3.6155391828533157, "grad_norm": 3.70403099273854, "learning_rate": 1.3943100386284685e-07, "loss": 0.8699, "step": 5398 }, { "epoch": 3.6162089752176825, "grad_norm": 2.86180721435989, "learning_rate": 1.3895015203898716e-07, "loss": 0.6854, "step": 5399 }, { "epoch": 3.6168787675820493, "grad_norm": 3.072679229302461, "learning_rate": 1.3847010709626907e-07, "loss": 0.9798, "step": 5400 }, { "epoch": 3.6175485599464166, "grad_norm": 3.4778278534873337, "learning_rate": 1.3799086919874554e-07, "loss": 0.6709, "step": 5401 }, { "epoch": 3.618218352310784, "grad_norm": 3.2174424921806453, "learning_rate": 1.3751243851019557e-07, "loss": 0.8243, "step": 5402 }, { "epoch": 3.6188881446751506, "grad_norm": 3.021284655586184, "learning_rate": 1.370348151941206e-07, "loss": 0.856, "step": 5403 }, { "epoch": 3.619557937039518, "grad_norm": 2.9459549155709577, "learning_rate": 1.3655799941374697e-07, "loss": 0.7514, "step": 5404 }, { "epoch": 3.6202277294038847, "grad_norm": 2.8960448146699758, "learning_rate": 1.360819913320252e-07, "loss": 0.8785, "step": 5405 }, { "epoch": 3.620897521768252, "grad_norm": 2.602216730704764, "learning_rate": 1.3560679111162932e-07, "loss": 0.8065, "step": 5406 }, { "epoch": 3.621567314132619, "grad_norm": 3.2935614542737164, "learning_rate": 1.351323989149575e-07, "loss": 0.9522, "step": 5407 }, { "epoch": 3.622237106496986, "grad_norm": 3.5825070948483186, "learning_rate": 1.3465881490413168e-07, "loss": 0.9285, "step": 5408 }, { "epoch": 3.622906898861353, "grad_norm": 3.210104044109568, "learning_rate": 1.341860392409977e-07, "loss": 0.9423, "step": 5409 }, { "epoch": 3.62357669122572, "grad_norm": 3.755698480327707, "learning_rate": 1.3371407208712523e-07, "loss": 0.9876, "step": 5410 }, { "epoch": 3.624246483590087, "grad_norm": 3.3355888006537198, "learning_rate": 1.332429136038074e-07, "loss": 0.6283, "step": 5411 }, { "epoch": 3.624916275954454, "grad_norm": 2.985163478733433, "learning_rate": 1.3277256395206106e-07, "loss": 0.7801, "step": 5412 }, { "epoch": 3.625586068318821, "grad_norm": 3.611595984538774, "learning_rate": 1.3230302329262674e-07, "loss": 0.7678, "step": 5413 }, { "epoch": 3.6262558606831883, "grad_norm": 3.448025137112942, "learning_rate": 1.3183429178596858e-07, "loss": 0.739, "step": 5414 }, { "epoch": 3.626925653047555, "grad_norm": 2.8577476600756246, "learning_rate": 1.3136636959227373e-07, "loss": 0.8063, "step": 5415 }, { "epoch": 3.6275954454119224, "grad_norm": 2.849788624009592, "learning_rate": 1.3089925687145345e-07, "loss": 0.8546, "step": 5416 }, { "epoch": 3.628265237776289, "grad_norm": 2.992427462154299, "learning_rate": 1.304329537831414e-07, "loss": 0.9035, "step": 5417 }, { "epoch": 3.6289350301406564, "grad_norm": 2.853610829259069, "learning_rate": 1.2996746048669602e-07, "loss": 0.901, "step": 5418 }, { "epoch": 3.6296048225050237, "grad_norm": 3.2154002075174897, "learning_rate": 1.2950277714119746e-07, "loss": 0.9028, "step": 5419 }, { "epoch": 3.6302746148693905, "grad_norm": 3.236251530299173, "learning_rate": 1.2903890390544989e-07, "loss": 0.9611, "step": 5420 }, { "epoch": 3.6309444072337573, "grad_norm": 3.7435100950616897, "learning_rate": 1.285758409379806e-07, "loss": 0.9906, "step": 5421 }, { "epoch": 3.6316141995981246, "grad_norm": 2.8879308419344762, "learning_rate": 1.281135883970397e-07, "loss": 0.6331, "step": 5422 }, { "epoch": 3.632283991962492, "grad_norm": 3.212945584312165, "learning_rate": 1.2765214644060109e-07, "loss": 0.8017, "step": 5423 }, { "epoch": 3.6329537843268587, "grad_norm": 3.371806640883716, "learning_rate": 1.271915152263603e-07, "loss": 0.9271, "step": 5424 }, { "epoch": 3.6336235766912255, "grad_norm": 2.918202462949644, "learning_rate": 1.2673169491173697e-07, "loss": 0.8632, "step": 5425 }, { "epoch": 3.6342933690555927, "grad_norm": 2.86059311633619, "learning_rate": 1.2627268565387262e-07, "loss": 0.8276, "step": 5426 }, { "epoch": 3.63496316141996, "grad_norm": 2.8328091643042947, "learning_rate": 1.2581448760963315e-07, "loss": 0.8433, "step": 5427 }, { "epoch": 3.635632953784327, "grad_norm": 2.9610448930044173, "learning_rate": 1.2535710093560583e-07, "loss": 0.7376, "step": 5428 }, { "epoch": 3.6363027461486936, "grad_norm": 3.2210616453001117, "learning_rate": 1.249005257881006e-07, "loss": 0.8343, "step": 5429 }, { "epoch": 3.636972538513061, "grad_norm": 2.6621783773330923, "learning_rate": 1.2444476232315094e-07, "loss": 0.6072, "step": 5430 }, { "epoch": 3.637642330877428, "grad_norm": 3.3891262779221885, "learning_rate": 1.239898106965129e-07, "loss": 0.9604, "step": 5431 }, { "epoch": 3.638312123241795, "grad_norm": 3.1330544592267717, "learning_rate": 1.2353567106366427e-07, "loss": 0.7878, "step": 5432 }, { "epoch": 3.6389819156061622, "grad_norm": 3.2014135412623395, "learning_rate": 1.2308234357980564e-07, "loss": 0.9416, "step": 5433 }, { "epoch": 3.639651707970529, "grad_norm": 2.7125491739262726, "learning_rate": 1.2262982839986032e-07, "loss": 0.8123, "step": 5434 }, { "epoch": 3.6403215003348963, "grad_norm": 2.9750856284848792, "learning_rate": 1.2217812567847432e-07, "loss": 0.7576, "step": 5435 }, { "epoch": 3.640991292699263, "grad_norm": 2.9915491605876423, "learning_rate": 1.2172723557001582e-07, "loss": 0.8898, "step": 5436 }, { "epoch": 3.6416610850636304, "grad_norm": 3.3163141938877794, "learning_rate": 1.2127715822857406e-07, "loss": 0.7967, "step": 5437 }, { "epoch": 3.642330877427997, "grad_norm": 3.1274204518305635, "learning_rate": 1.2082789380796155e-07, "loss": 0.9143, "step": 5438 }, { "epoch": 3.6430006697923645, "grad_norm": 3.1537116275205546, "learning_rate": 1.2037944246171386e-07, "loss": 0.7482, "step": 5439 }, { "epoch": 3.6436704621567313, "grad_norm": 3.314588746199295, "learning_rate": 1.1993180434308775e-07, "loss": 0.9287, "step": 5440 }, { "epoch": 3.6443402545210986, "grad_norm": 2.9786378052181472, "learning_rate": 1.1948497960506122e-07, "loss": 0.8495, "step": 5441 }, { "epoch": 3.6450100468854654, "grad_norm": 3.1535391560675263, "learning_rate": 1.1903896840033513e-07, "loss": 0.8712, "step": 5442 }, { "epoch": 3.6456798392498326, "grad_norm": 2.898504679629912, "learning_rate": 1.185937708813334e-07, "loss": 0.8442, "step": 5443 }, { "epoch": 3.6463496316141994, "grad_norm": 3.1325938635294532, "learning_rate": 1.1814938720020069e-07, "loss": 0.9555, "step": 5444 }, { "epoch": 3.6470194239785667, "grad_norm": 3.2932527610626687, "learning_rate": 1.1770581750880305e-07, "loss": 0.798, "step": 5445 }, { "epoch": 3.6476892163429335, "grad_norm": 2.9464452133359273, "learning_rate": 1.1726306195872866e-07, "loss": 0.9666, "step": 5446 }, { "epoch": 3.648359008707301, "grad_norm": 2.955983666434013, "learning_rate": 1.1682112070128871e-07, "loss": 0.8193, "step": 5447 }, { "epoch": 3.649028801071668, "grad_norm": 2.7832997441089113, "learning_rate": 1.1637999388751542e-07, "loss": 0.914, "step": 5448 }, { "epoch": 3.649698593436035, "grad_norm": 3.5597141093230813, "learning_rate": 1.1593968166816128e-07, "loss": 0.7044, "step": 5449 }, { "epoch": 3.6503683858004017, "grad_norm": 2.988516723253611, "learning_rate": 1.1550018419370201e-07, "loss": 0.9433, "step": 5450 }, { "epoch": 3.651038178164769, "grad_norm": 2.6627877305740184, "learning_rate": 1.1506150161433499e-07, "loss": 0.8537, "step": 5451 }, { "epoch": 3.651707970529136, "grad_norm": 3.707928184772422, "learning_rate": 1.1462363407997834e-07, "loss": 0.7885, "step": 5452 }, { "epoch": 3.652377762893503, "grad_norm": 3.114449153724966, "learning_rate": 1.1418658174027125e-07, "loss": 0.8688, "step": 5453 }, { "epoch": 3.65304755525787, "grad_norm": 3.100656066195091, "learning_rate": 1.1375034474457563e-07, "loss": 0.8779, "step": 5454 }, { "epoch": 3.653717347622237, "grad_norm": 2.922780551084493, "learning_rate": 1.133149232419739e-07, "loss": 0.8442, "step": 5455 }, { "epoch": 3.6543871399866044, "grad_norm": 3.0004574978010305, "learning_rate": 1.1288031738127008e-07, "loss": 0.9678, "step": 5456 }, { "epoch": 3.655056932350971, "grad_norm": 2.9680630459990613, "learning_rate": 1.1244652731098871e-07, "loss": 0.8641, "step": 5457 }, { "epoch": 3.655726724715338, "grad_norm": 3.102399374390752, "learning_rate": 1.1201355317937673e-07, "loss": 0.8692, "step": 5458 }, { "epoch": 3.6563965170797053, "grad_norm": 3.4292110906633155, "learning_rate": 1.1158139513440163e-07, "loss": 1.0055, "step": 5459 }, { "epoch": 3.6570663094440725, "grad_norm": 2.75280892690931, "learning_rate": 1.1115005332375195e-07, "loss": 0.7527, "step": 5460 }, { "epoch": 3.6577361018084393, "grad_norm": 2.950712870837497, "learning_rate": 1.107195278948367e-07, "loss": 0.8881, "step": 5461 }, { "epoch": 3.6584058941728066, "grad_norm": 3.1443215836204144, "learning_rate": 1.1028981899478763e-07, "loss": 0.9707, "step": 5462 }, { "epoch": 3.6590756865371734, "grad_norm": 2.9548816024741833, "learning_rate": 1.098609267704559e-07, "loss": 0.8165, "step": 5463 }, { "epoch": 3.6597454789015407, "grad_norm": 3.069057557796588, "learning_rate": 1.0943285136841398e-07, "loss": 0.8337, "step": 5464 }, { "epoch": 3.6604152712659075, "grad_norm": 3.7663077118495827, "learning_rate": 1.090055929349551e-07, "loss": 0.7918, "step": 5465 }, { "epoch": 3.6610850636302747, "grad_norm": 3.2621619975607374, "learning_rate": 1.0857915161609361e-07, "loss": 0.9536, "step": 5466 }, { "epoch": 3.6617548559946416, "grad_norm": 3.2336083182300412, "learning_rate": 1.0815352755756458e-07, "loss": 0.9463, "step": 5467 }, { "epoch": 3.662424648359009, "grad_norm": 3.051802344295739, "learning_rate": 1.0772872090482389e-07, "loss": 0.7473, "step": 5468 }, { "epoch": 3.6630944407233756, "grad_norm": 3.4368129474656435, "learning_rate": 1.0730473180304679e-07, "loss": 0.9963, "step": 5469 }, { "epoch": 3.663764233087743, "grad_norm": 3.1054936323570583, "learning_rate": 1.0688156039713127e-07, "loss": 0.9484, "step": 5470 }, { "epoch": 3.6644340254521097, "grad_norm": 2.8460787538087033, "learning_rate": 1.0645920683169442e-07, "loss": 0.8632, "step": 5471 }, { "epoch": 3.665103817816477, "grad_norm": 3.694145794940387, "learning_rate": 1.0603767125107412e-07, "loss": 0.805, "step": 5472 }, { "epoch": 3.665773610180844, "grad_norm": 3.2167580256760697, "learning_rate": 1.0561695379932874e-07, "loss": 0.8597, "step": 5473 }, { "epoch": 3.666443402545211, "grad_norm": 3.8275285607546077, "learning_rate": 1.0519705462023744e-07, "loss": 0.7994, "step": 5474 }, { "epoch": 3.667113194909578, "grad_norm": 3.397546643703304, "learning_rate": 1.0477797385729905e-07, "loss": 0.8046, "step": 5475 }, { "epoch": 3.667782987273945, "grad_norm": 2.9719755393699994, "learning_rate": 1.0435971165373316e-07, "loss": 0.9171, "step": 5476 }, { "epoch": 3.6684527796383124, "grad_norm": 2.958319240950184, "learning_rate": 1.039422681524796e-07, "loss": 0.8332, "step": 5477 }, { "epoch": 3.669122572002679, "grad_norm": 3.217476710378833, "learning_rate": 1.0352564349619842e-07, "loss": 0.8771, "step": 5478 }, { "epoch": 3.669792364367046, "grad_norm": 3.0200551220661755, "learning_rate": 1.0310983782726936e-07, "loss": 0.804, "step": 5479 }, { "epoch": 3.6704621567314133, "grad_norm": 3.39785255578848, "learning_rate": 1.0269485128779316e-07, "loss": 0.925, "step": 5480 }, { "epoch": 3.6711319490957806, "grad_norm": 2.966284034434669, "learning_rate": 1.0228068401958973e-07, "loss": 0.8645, "step": 5481 }, { "epoch": 3.6718017414601474, "grad_norm": 2.9221105978067947, "learning_rate": 1.0186733616419974e-07, "loss": 0.8698, "step": 5482 }, { "epoch": 3.672471533824514, "grad_norm": 3.284179384229182, "learning_rate": 1.0145480786288326e-07, "loss": 0.7565, "step": 5483 }, { "epoch": 3.6731413261888815, "grad_norm": 3.0323021642283594, "learning_rate": 1.0104309925662053e-07, "loss": 0.885, "step": 5484 }, { "epoch": 3.6738111185532487, "grad_norm": 2.865046663192907, "learning_rate": 1.0063221048611155e-07, "loss": 0.763, "step": 5485 }, { "epoch": 3.6744809109176155, "grad_norm": 3.2375314819460206, "learning_rate": 1.0022214169177647e-07, "loss": 0.9454, "step": 5486 }, { "epoch": 3.6751507032819823, "grad_norm": 3.0804404317518723, "learning_rate": 9.981289301375458e-08, "loss": 0.8007, "step": 5487 }, { "epoch": 3.6758204956463496, "grad_norm": 3.1374842268465684, "learning_rate": 9.940446459190567e-08, "loss": 0.9037, "step": 5488 }, { "epoch": 3.676490288010717, "grad_norm": 3.098639342609978, "learning_rate": 9.899685656580866e-08, "loss": 0.9286, "step": 5489 }, { "epoch": 3.6771600803750837, "grad_norm": 2.750522520745243, "learning_rate": 9.859006907476237e-08, "loss": 0.7557, "step": 5490 }, { "epoch": 3.6778298727394505, "grad_norm": 3.03410525018198, "learning_rate": 9.818410225778507e-08, "loss": 0.9206, "step": 5491 }, { "epoch": 3.6784996651038178, "grad_norm": 3.0828056859856097, "learning_rate": 9.777895625361439e-08, "loss": 0.8601, "step": 5492 }, { "epoch": 3.679169457468185, "grad_norm": 2.956707155381972, "learning_rate": 9.737463120070817e-08, "loss": 0.7586, "step": 5493 }, { "epoch": 3.679839249832552, "grad_norm": 2.809861986484399, "learning_rate": 9.697112723724256e-08, "loss": 0.7977, "step": 5494 }, { "epoch": 3.680509042196919, "grad_norm": 3.304569763285542, "learning_rate": 9.656844450111419e-08, "loss": 0.9573, "step": 5495 }, { "epoch": 3.681178834561286, "grad_norm": 3.3250838856457845, "learning_rate": 9.616658312993855e-08, "loss": 0.9388, "step": 5496 }, { "epoch": 3.681848626925653, "grad_norm": 3.7552766168973304, "learning_rate": 9.57655432610502e-08, "loss": 0.9282, "step": 5497 }, { "epoch": 3.68251841929002, "grad_norm": 2.7650313752769042, "learning_rate": 9.536532503150341e-08, "loss": 0.7108, "step": 5498 }, { "epoch": 3.6831882116543873, "grad_norm": 2.948927778479058, "learning_rate": 9.49659285780713e-08, "loss": 0.8839, "step": 5499 }, { "epoch": 3.683858004018754, "grad_norm": 3.9752257523759194, "learning_rate": 9.45673540372466e-08, "loss": 0.9764, "step": 5500 }, { "epoch": 3.6845277963831213, "grad_norm": 3.326103731352437, "learning_rate": 9.416960154524062e-08, "loss": 0.8102, "step": 5501 }, { "epoch": 3.685197588747488, "grad_norm": 3.432538938703598, "learning_rate": 9.377267123798411e-08, "loss": 0.8016, "step": 5502 }, { "epoch": 3.6858673811118554, "grad_norm": 2.5180849388899005, "learning_rate": 9.33765632511266e-08, "loss": 0.8, "step": 5503 }, { "epoch": 3.6865371734762222, "grad_norm": 3.251357758433833, "learning_rate": 9.298127772003701e-08, "loss": 0.9438, "step": 5504 }, { "epoch": 3.6872069658405895, "grad_norm": 3.3589827177574136, "learning_rate": 9.258681477980231e-08, "loss": 0.9932, "step": 5505 }, { "epoch": 3.6878767582049563, "grad_norm": 3.148804231664732, "learning_rate": 9.219317456522963e-08, "loss": 0.7253, "step": 5506 }, { "epoch": 3.6885465505693236, "grad_norm": 3.148465567884084, "learning_rate": 9.180035721084363e-08, "loss": 1.0145, "step": 5507 }, { "epoch": 3.6892163429336904, "grad_norm": 2.83738725548807, "learning_rate": 9.140836285088884e-08, "loss": 0.8138, "step": 5508 }, { "epoch": 3.6898861352980576, "grad_norm": 2.8475225382508205, "learning_rate": 9.101719161932753e-08, "loss": 0.8293, "step": 5509 }, { "epoch": 3.690555927662425, "grad_norm": 3.121811912916505, "learning_rate": 9.062684364984226e-08, "loss": 0.7727, "step": 5510 }, { "epoch": 3.6912257200267917, "grad_norm": 3.555728070724802, "learning_rate": 9.023731907583182e-08, "loss": 0.9812, "step": 5511 }, { "epoch": 3.6918955123911585, "grad_norm": 2.965300619392912, "learning_rate": 8.98486180304156e-08, "loss": 0.7878, "step": 5512 }, { "epoch": 3.692565304755526, "grad_norm": 3.106851025765158, "learning_rate": 8.946074064643124e-08, "loss": 0.8167, "step": 5513 }, { "epoch": 3.693235097119893, "grad_norm": 3.110641109828901, "learning_rate": 8.907368705643432e-08, "loss": 0.8956, "step": 5514 }, { "epoch": 3.69390488948426, "grad_norm": 2.8027428486040744, "learning_rate": 8.868745739269907e-08, "loss": 0.821, "step": 5515 }, { "epoch": 3.6945746818486267, "grad_norm": 2.726852807763832, "learning_rate": 8.830205178721768e-08, "loss": 0.7347, "step": 5516 }, { "epoch": 3.695244474212994, "grad_norm": 4.3257361294572165, "learning_rate": 8.791747037170228e-08, "loss": 0.7423, "step": 5517 }, { "epoch": 3.695914266577361, "grad_norm": 2.7845973279489002, "learning_rate": 8.753371327758248e-08, "loss": 0.8196, "step": 5518 }, { "epoch": 3.696584058941728, "grad_norm": 3.4710962121983644, "learning_rate": 8.715078063600474e-08, "loss": 0.8782, "step": 5519 }, { "epoch": 3.697253851306095, "grad_norm": 3.4295432537719814, "learning_rate": 8.676867257783578e-08, "loss": 0.9697, "step": 5520 }, { "epoch": 3.697923643670462, "grad_norm": 2.7164290126042845, "learning_rate": 8.638738923365974e-08, "loss": 0.8581, "step": 5521 }, { "epoch": 3.6985934360348294, "grad_norm": 3.096319794932549, "learning_rate": 8.600693073377963e-08, "loss": 0.8173, "step": 5522 }, { "epoch": 3.699263228399196, "grad_norm": 3.013199991509023, "learning_rate": 8.562729720821478e-08, "loss": 0.7462, "step": 5523 }, { "epoch": 3.6999330207635635, "grad_norm": 2.686873555364831, "learning_rate": 8.524848878670389e-08, "loss": 0.8149, "step": 5524 }, { "epoch": 3.7006028131279303, "grad_norm": 3.1348337004112974, "learning_rate": 8.487050559870425e-08, "loss": 0.7918, "step": 5525 }, { "epoch": 3.7012726054922975, "grad_norm": 2.8601524539833463, "learning_rate": 8.449334777339003e-08, "loss": 0.8831, "step": 5526 }, { "epoch": 3.7019423978566643, "grad_norm": 3.3654703175219938, "learning_rate": 8.411701543965367e-08, "loss": 0.9031, "step": 5527 }, { "epoch": 3.7026121902210316, "grad_norm": 2.912972236493136, "learning_rate": 8.374150872610482e-08, "loss": 0.6331, "step": 5528 }, { "epoch": 3.7032819825853984, "grad_norm": 3.3330551850280328, "learning_rate": 8.336682776107302e-08, "loss": 0.918, "step": 5529 }, { "epoch": 3.7039517749497657, "grad_norm": 3.1811406225742713, "learning_rate": 8.299297267260365e-08, "loss": 1.047, "step": 5530 }, { "epoch": 3.7046215673141325, "grad_norm": 2.8369166261229797, "learning_rate": 8.261994358846031e-08, "loss": 0.8799, "step": 5531 }, { "epoch": 3.7052913596784998, "grad_norm": 2.8075069270239146, "learning_rate": 8.22477406361244e-08, "loss": 0.8919, "step": 5532 }, { "epoch": 3.7059611520428666, "grad_norm": 2.7704388908075157, "learning_rate": 8.187636394279524e-08, "loss": 0.7989, "step": 5533 }, { "epoch": 3.706630944407234, "grad_norm": 3.111983477253092, "learning_rate": 8.150581363539022e-08, "loss": 0.7927, "step": 5534 }, { "epoch": 3.7073007367716007, "grad_norm": 2.883264852160906, "learning_rate": 8.113608984054278e-08, "loss": 0.8464, "step": 5535 }, { "epoch": 3.707970529135968, "grad_norm": 2.758113430093421, "learning_rate": 8.076719268460487e-08, "loss": 0.8162, "step": 5536 }, { "epoch": 3.7086403215003347, "grad_norm": 3.0940066895952576, "learning_rate": 8.039912229364677e-08, "loss": 0.8518, "step": 5537 }, { "epoch": 3.709310113864702, "grad_norm": 3.1827372331479546, "learning_rate": 8.003187879345481e-08, "loss": 0.846, "step": 5538 }, { "epoch": 3.7099799062290693, "grad_norm": 3.6521495077746056, "learning_rate": 7.966546230953276e-08, "loss": 0.8937, "step": 5539 }, { "epoch": 3.710649698593436, "grad_norm": 3.14607102635032, "learning_rate": 7.929987296710295e-08, "loss": 0.9944, "step": 5540 }, { "epoch": 3.711319490957803, "grad_norm": 3.210508231601047, "learning_rate": 7.893511089110434e-08, "loss": 0.9823, "step": 5541 }, { "epoch": 3.71198928332217, "grad_norm": 2.8422752029883465, "learning_rate": 7.857117620619331e-08, "loss": 0.7812, "step": 5542 }, { "epoch": 3.7126590756865374, "grad_norm": 3.1200597052766934, "learning_rate": 7.820806903674232e-08, "loss": 0.8726, "step": 5543 }, { "epoch": 3.7133288680509042, "grad_norm": 2.923545730088978, "learning_rate": 7.784578950684324e-08, "loss": 0.8133, "step": 5544 }, { "epoch": 3.713998660415271, "grad_norm": 3.0469310406308536, "learning_rate": 7.74843377403034e-08, "loss": 0.9447, "step": 5545 }, { "epoch": 3.7146684527796383, "grad_norm": 2.8346036454698593, "learning_rate": 7.712371386064788e-08, "loss": 1.0559, "step": 5546 }, { "epoch": 3.7153382451440056, "grad_norm": 3.115065742550323, "learning_rate": 7.676391799111893e-08, "loss": 0.9502, "step": 5547 }, { "epoch": 3.7160080375083724, "grad_norm": 3.291963961549783, "learning_rate": 7.640495025467515e-08, "loss": 0.8588, "step": 5548 }, { "epoch": 3.716677829872739, "grad_norm": 3.305874850508865, "learning_rate": 7.604681077399312e-08, "loss": 0.9238, "step": 5549 }, { "epoch": 3.7173476222371065, "grad_norm": 3.2455586171806226, "learning_rate": 7.56894996714655e-08, "loss": 0.9698, "step": 5550 }, { "epoch": 3.7180174146014737, "grad_norm": 4.888522189953705, "learning_rate": 7.53330170692021e-08, "loss": 0.9343, "step": 5551 }, { "epoch": 3.7186872069658405, "grad_norm": 2.660452493929928, "learning_rate": 7.49773630890302e-08, "loss": 0.8186, "step": 5552 }, { "epoch": 3.7193569993302074, "grad_norm": 3.673722617125322, "learning_rate": 7.462253785249285e-08, "loss": 0.9104, "step": 5553 }, { "epoch": 3.7200267916945746, "grad_norm": 3.069802699300511, "learning_rate": 7.426854148085055e-08, "loss": 0.8472, "step": 5554 }, { "epoch": 3.720696584058942, "grad_norm": 3.8408246817101275, "learning_rate": 7.391537409508071e-08, "loss": 0.8171, "step": 5555 }, { "epoch": 3.7213663764233087, "grad_norm": 2.917342439923884, "learning_rate": 7.356303581587704e-08, "loss": 0.8041, "step": 5556 }, { "epoch": 3.722036168787676, "grad_norm": 4.056467729184028, "learning_rate": 7.32115267636499e-08, "loss": 0.9021, "step": 5557 }, { "epoch": 3.7227059611520428, "grad_norm": 3.066800321581996, "learning_rate": 7.286084705852625e-08, "loss": 0.9518, "step": 5558 }, { "epoch": 3.72337575351641, "grad_norm": 2.897118114711809, "learning_rate": 7.251099682034996e-08, "loss": 0.8336, "step": 5559 }, { "epoch": 3.724045545880777, "grad_norm": 3.194949663132412, "learning_rate": 7.21619761686812e-08, "loss": 0.8074, "step": 5560 }, { "epoch": 3.724715338245144, "grad_norm": 2.9445723498312755, "learning_rate": 7.181378522279653e-08, "loss": 0.9713, "step": 5561 }, { "epoch": 3.725385130609511, "grad_norm": 2.991196077071294, "learning_rate": 7.146642410168907e-08, "loss": 0.7478, "step": 5562 }, { "epoch": 3.726054922973878, "grad_norm": 3.9861729383092857, "learning_rate": 7.111989292406862e-08, "loss": 0.8146, "step": 5563 }, { "epoch": 3.726724715338245, "grad_norm": 2.800071304891211, "learning_rate": 7.07741918083607e-08, "loss": 0.8208, "step": 5564 }, { "epoch": 3.7273945077026123, "grad_norm": 3.2079461041746273, "learning_rate": 7.042932087270776e-08, "loss": 0.8943, "step": 5565 }, { "epoch": 3.728064300066979, "grad_norm": 3.05438207597744, "learning_rate": 7.00852802349683e-08, "loss": 0.7527, "step": 5566 }, { "epoch": 3.7287340924313463, "grad_norm": 2.8383510697718104, "learning_rate": 6.974207001271716e-08, "loss": 0.7691, "step": 5567 }, { "epoch": 3.729403884795713, "grad_norm": 3.200965376636467, "learning_rate": 6.939969032324523e-08, "loss": 0.8442, "step": 5568 }, { "epoch": 3.7300736771600804, "grad_norm": 2.916566819673214, "learning_rate": 6.905814128355975e-08, "loss": 0.9733, "step": 5569 }, { "epoch": 3.7307434695244472, "grad_norm": 3.498354521040086, "learning_rate": 6.871742301038403e-08, "loss": 0.964, "step": 5570 }, { "epoch": 3.7314132618888145, "grad_norm": 3.018769662952501, "learning_rate": 6.837753562015714e-08, "loss": 0.7573, "step": 5571 }, { "epoch": 3.7320830542531818, "grad_norm": 3.043659532784094, "learning_rate": 6.803847922903478e-08, "loss": 0.7617, "step": 5572 }, { "epoch": 3.7327528466175486, "grad_norm": 3.333325193045494, "learning_rate": 6.770025395288842e-08, "loss": 0.8086, "step": 5573 }, { "epoch": 3.7334226389819154, "grad_norm": 3.103560056570463, "learning_rate": 6.736285990730507e-08, "loss": 0.8555, "step": 5574 }, { "epoch": 3.7340924313462827, "grad_norm": 2.8418362943619453, "learning_rate": 6.702629720758857e-08, "loss": 0.8374, "step": 5575 }, { "epoch": 3.73476222371065, "grad_norm": 3.105173340549629, "learning_rate": 6.669056596875778e-08, "loss": 0.9225, "step": 5576 }, { "epoch": 3.7354320160750167, "grad_norm": 3.2170106677282355, "learning_rate": 6.635566630554785e-08, "loss": 0.9916, "step": 5577 }, { "epoch": 3.7361018084393836, "grad_norm": 2.9170696806939644, "learning_rate": 6.602159833240973e-08, "loss": 0.8542, "step": 5578 }, { "epoch": 3.736771600803751, "grad_norm": 2.842752565191429, "learning_rate": 6.56883621635096e-08, "loss": 0.8633, "step": 5579 }, { "epoch": 3.737441393168118, "grad_norm": 3.1428848571305252, "learning_rate": 6.535595791273108e-08, "loss": 0.9186, "step": 5580 }, { "epoch": 3.738111185532485, "grad_norm": 2.870215250698304, "learning_rate": 6.502438569367081e-08, "loss": 0.8207, "step": 5581 }, { "epoch": 3.7387809778968517, "grad_norm": 2.844066357352238, "learning_rate": 6.469364561964315e-08, "loss": 0.8018, "step": 5582 }, { "epoch": 3.739450770261219, "grad_norm": 3.1522029275734917, "learning_rate": 6.436373780367739e-08, "loss": 0.9069, "step": 5583 }, { "epoch": 3.7401205626255862, "grad_norm": 3.6391411411852106, "learning_rate": 6.40346623585189e-08, "loss": 0.9283, "step": 5584 }, { "epoch": 3.740790354989953, "grad_norm": 3.2386972009819552, "learning_rate": 6.370641939662748e-08, "loss": 0.6759, "step": 5585 }, { "epoch": 3.7414601473543203, "grad_norm": 3.145928018061326, "learning_rate": 6.337900903017946e-08, "loss": 0.961, "step": 5586 }, { "epoch": 3.742129939718687, "grad_norm": 2.536010721953197, "learning_rate": 6.305243137106564e-08, "loss": 0.6064, "step": 5587 }, { "epoch": 3.7427997320830544, "grad_norm": 2.807179915210428, "learning_rate": 6.272668653089426e-08, "loss": 0.8691, "step": 5588 }, { "epoch": 3.743469524447421, "grad_norm": 3.1564374560718855, "learning_rate": 6.240177462098624e-08, "loss": 0.7715, "step": 5589 }, { "epoch": 3.7441393168117885, "grad_norm": 3.1745400257703587, "learning_rate": 6.20776957523797e-08, "loss": 0.8257, "step": 5590 }, { "epoch": 3.7448091091761553, "grad_norm": 3.1532454130118666, "learning_rate": 6.17544500358272e-08, "loss": 0.7786, "step": 5591 }, { "epoch": 3.7454789015405225, "grad_norm": 2.672824813374942, "learning_rate": 6.143203758179783e-08, "loss": 0.8634, "step": 5592 }, { "epoch": 3.7461486939048894, "grad_norm": 3.707030107524006, "learning_rate": 6.111045850047404e-08, "loss": 0.9077, "step": 5593 }, { "epoch": 3.7468184862692566, "grad_norm": 3.4677382549974833, "learning_rate": 6.078971290175484e-08, "loss": 0.9009, "step": 5594 }, { "epoch": 3.7474882786336234, "grad_norm": 3.025872508989665, "learning_rate": 6.046980089525367e-08, "loss": 0.982, "step": 5595 }, { "epoch": 3.7481580709979907, "grad_norm": 2.939029079183772, "learning_rate": 6.015072259030003e-08, "loss": 0.8329, "step": 5596 }, { "epoch": 3.7488278633623575, "grad_norm": 3.8899231594025494, "learning_rate": 5.983247809593751e-08, "loss": 0.8389, "step": 5597 }, { "epoch": 3.7494976557267248, "grad_norm": 3.3664781499454683, "learning_rate": 5.951506752092523e-08, "loss": 0.8131, "step": 5598 }, { "epoch": 3.7501674480910916, "grad_norm": 3.2417827106041273, "learning_rate": 5.91984909737367e-08, "loss": 0.8589, "step": 5599 }, { "epoch": 3.750837240455459, "grad_norm": 3.184897645688921, "learning_rate": 5.8882748562562045e-08, "loss": 0.8992, "step": 5600 }, { "epoch": 3.751507032819826, "grad_norm": 2.8414262092525764, "learning_rate": 5.856784039530439e-08, "loss": 0.7705, "step": 5601 }, { "epoch": 3.752176825184193, "grad_norm": 3.1037909111509254, "learning_rate": 5.825376657958237e-08, "loss": 0.9021, "step": 5602 }, { "epoch": 3.7528466175485597, "grad_norm": 3.165460122842699, "learning_rate": 5.794052722273041e-08, "loss": 0.8734, "step": 5603 }, { "epoch": 3.753516409912927, "grad_norm": 3.1727291345727404, "learning_rate": 5.762812243179705e-08, "loss": 0.9418, "step": 5604 }, { "epoch": 3.7541862022772943, "grad_norm": 3.2649690277721137, "learning_rate": 5.731655231354521e-08, "loss": 0.8821, "step": 5605 }, { "epoch": 3.754855994641661, "grad_norm": 3.848838435265384, "learning_rate": 5.700581697445251e-08, "loss": 0.8904, "step": 5606 }, { "epoch": 3.755525787006028, "grad_norm": 4.794477267925431, "learning_rate": 5.6695916520712604e-08, "loss": 0.9208, "step": 5607 }, { "epoch": 3.756195579370395, "grad_norm": 3.044304758913152, "learning_rate": 5.6386851058232995e-08, "loss": 0.8582, "step": 5608 }, { "epoch": 3.7568653717347624, "grad_norm": 2.9007770333509093, "learning_rate": 5.607862069263531e-08, "loss": 0.8889, "step": 5609 }, { "epoch": 3.7575351640991292, "grad_norm": 2.950488305441694, "learning_rate": 5.5771225529255844e-08, "loss": 0.9022, "step": 5610 }, { "epoch": 3.758204956463496, "grad_norm": 2.8819350419611, "learning_rate": 5.546466567314695e-08, "loss": 0.9511, "step": 5611 }, { "epoch": 3.7588747488278633, "grad_norm": 3.0532801515080514, "learning_rate": 5.515894122907428e-08, "loss": 0.7631, "step": 5612 }, { "epoch": 3.7595445411922306, "grad_norm": 3.6820335566761893, "learning_rate": 5.4854052301517576e-08, "loss": 0.8226, "step": 5613 }, { "epoch": 3.7602143335565974, "grad_norm": 5.284498162709322, "learning_rate": 5.454999899467184e-08, "loss": 0.684, "step": 5614 }, { "epoch": 3.7608841259209647, "grad_norm": 2.7061689821715373, "learning_rate": 5.4246781412446715e-08, "loss": 0.9557, "step": 5615 }, { "epoch": 3.7615539182853315, "grad_norm": 3.383244582612517, "learning_rate": 5.394439965846543e-08, "loss": 0.8346, "step": 5616 }, { "epoch": 3.7622237106496987, "grad_norm": 3.0316861039187866, "learning_rate": 5.3642853836066434e-08, "loss": 0.6849, "step": 5617 }, { "epoch": 3.7628935030140656, "grad_norm": 2.9918714988679382, "learning_rate": 5.3342144048300904e-08, "loss": 0.7492, "step": 5618 }, { "epoch": 3.763563295378433, "grad_norm": 2.9722905307901795, "learning_rate": 5.304227039793691e-08, "loss": 0.8203, "step": 5619 }, { "epoch": 3.7642330877427996, "grad_norm": 2.917682190772594, "learning_rate": 5.2743232987454154e-08, "loss": 0.9605, "step": 5620 }, { "epoch": 3.764902880107167, "grad_norm": 3.080254779857474, "learning_rate": 5.244503191904865e-08, "loss": 0.9769, "step": 5621 }, { "epoch": 3.7655726724715337, "grad_norm": 3.0434500330439684, "learning_rate": 5.214766729462861e-08, "loss": 0.8527, "step": 5622 }, { "epoch": 3.766242464835901, "grad_norm": 3.2279293915749783, "learning_rate": 5.1851139215818304e-08, "loss": 0.8771, "step": 5623 }, { "epoch": 3.766912257200268, "grad_norm": 2.755054658045152, "learning_rate": 5.1555447783954994e-08, "loss": 0.8433, "step": 5624 }, { "epoch": 3.767582049564635, "grad_norm": 3.0793015593835498, "learning_rate": 5.126059310009007e-08, "loss": 0.9567, "step": 5625 }, { "epoch": 3.768251841929002, "grad_norm": 3.041405315687881, "learning_rate": 5.0966575264989046e-08, "loss": 0.9107, "step": 5626 }, { "epoch": 3.768921634293369, "grad_norm": 3.1636336349222445, "learning_rate": 5.06733943791321e-08, "loss": 0.8633, "step": 5627 }, { "epoch": 3.769591426657736, "grad_norm": 2.7006982645042124, "learning_rate": 5.0381050542712415e-08, "loss": 0.7721, "step": 5628 }, { "epoch": 3.770261219022103, "grad_norm": 2.9869921908911197, "learning_rate": 5.008954385563786e-08, "loss": 1.0305, "step": 5629 }, { "epoch": 3.7709310113864705, "grad_norm": 3.02879824858855, "learning_rate": 4.9798874417529844e-08, "loss": 0.876, "step": 5630 }, { "epoch": 3.7716008037508373, "grad_norm": 2.860756638351461, "learning_rate": 4.950904232772336e-08, "loss": 0.8569, "step": 5631 }, { "epoch": 3.772270596115204, "grad_norm": 3.122491849425746, "learning_rate": 4.922004768526778e-08, "loss": 0.8664, "step": 5632 }, { "epoch": 3.7729403884795714, "grad_norm": 3.025861147992542, "learning_rate": 4.8931890588926334e-08, "loss": 0.845, "step": 5633 }, { "epoch": 3.7736101808439386, "grad_norm": 3.6514866340413814, "learning_rate": 4.864457113717525e-08, "loss": 0.9061, "step": 5634 }, { "epoch": 3.7742799732083054, "grad_norm": 3.600129830461874, "learning_rate": 4.835808942820569e-08, "loss": 0.9337, "step": 5635 }, { "epoch": 3.7749497655726723, "grad_norm": 3.189903149556754, "learning_rate": 4.80724455599213e-08, "loss": 0.9133, "step": 5636 }, { "epoch": 3.7756195579370395, "grad_norm": 2.7524367765949025, "learning_rate": 4.7787639629940095e-08, "loss": 0.8435, "step": 5637 }, { "epoch": 3.7762893503014068, "grad_norm": 3.224364982083639, "learning_rate": 4.750367173559395e-08, "loss": 0.805, "step": 5638 }, { "epoch": 3.7769591426657736, "grad_norm": 4.2818096020478915, "learning_rate": 4.722054197392745e-08, "loss": 0.7072, "step": 5639 }, { "epoch": 3.7776289350301404, "grad_norm": 2.6264436994950553, "learning_rate": 4.6938250441699584e-08, "loss": 0.8433, "step": 5640 }, { "epoch": 3.7782987273945077, "grad_norm": 2.951211752957764, "learning_rate": 4.665679723538235e-08, "loss": 0.9245, "step": 5641 }, { "epoch": 3.778968519758875, "grad_norm": 3.0700814534904586, "learning_rate": 4.63761824511616e-08, "loss": 0.7169, "step": 5642 }, { "epoch": 3.7796383121232418, "grad_norm": 3.0275086463254444, "learning_rate": 4.609640618493644e-08, "loss": 0.8216, "step": 5643 }, { "epoch": 3.7803081044876086, "grad_norm": 3.1004936249409427, "learning_rate": 4.581746853231983e-08, "loss": 0.9001, "step": 5644 }, { "epoch": 3.780977896851976, "grad_norm": 3.3227067443492424, "learning_rate": 4.553936958863747e-08, "loss": 0.836, "step": 5645 }, { "epoch": 3.781647689216343, "grad_norm": 2.935394720844692, "learning_rate": 4.526210944892862e-08, "loss": 0.7894, "step": 5646 }, { "epoch": 3.78231748158071, "grad_norm": 3.4893494726028957, "learning_rate": 4.498568820794663e-08, "loss": 0.762, "step": 5647 }, { "epoch": 3.782987273945077, "grad_norm": 2.8325712619831287, "learning_rate": 4.471010596015707e-08, "loss": 0.9416, "step": 5648 }, { "epoch": 3.783657066309444, "grad_norm": 3.1337457779359505, "learning_rate": 4.443536279973931e-08, "loss": 0.8603, "step": 5649 }, { "epoch": 3.7843268586738112, "grad_norm": 3.0858915151237376, "learning_rate": 4.416145882058603e-08, "loss": 0.9871, "step": 5650 }, { "epoch": 3.784996651038178, "grad_norm": 3.1709210440719025, "learning_rate": 4.388839411630319e-08, "loss": 0.9087, "step": 5651 }, { "epoch": 3.7856664434025453, "grad_norm": 2.8258122315423275, "learning_rate": 4.361616878020974e-08, "loss": 0.7318, "step": 5652 }, { "epoch": 3.786336235766912, "grad_norm": 2.9156600337427516, "learning_rate": 4.3344782905337393e-08, "loss": 0.7946, "step": 5653 }, { "epoch": 3.7870060281312794, "grad_norm": 2.909005815361952, "learning_rate": 4.3074236584431937e-08, "loss": 0.8881, "step": 5654 }, { "epoch": 3.787675820495646, "grad_norm": 3.2830190346209, "learning_rate": 4.280452990995137e-08, "loss": 0.8422, "step": 5655 }, { "epoch": 3.7883456128600135, "grad_norm": 2.971868483545581, "learning_rate": 4.253566297406725e-08, "loss": 0.782, "step": 5656 }, { "epoch": 3.7890154052243803, "grad_norm": 2.8444863580637185, "learning_rate": 4.226763586866384e-08, "loss": 0.9285, "step": 5657 }, { "epoch": 3.7896851975887476, "grad_norm": 3.2755784718630383, "learning_rate": 4.2000448685338423e-08, "loss": 0.8708, "step": 5658 }, { "epoch": 3.7903549899531144, "grad_norm": 3.051677935366405, "learning_rate": 4.173410151540186e-08, "loss": 0.9949, "step": 5659 }, { "epoch": 3.7910247823174816, "grad_norm": 4.033872398938437, "learning_rate": 4.146859444987689e-08, "loss": 0.8228, "step": 5660 }, { "epoch": 3.7916945746818485, "grad_norm": 3.2262173558093985, "learning_rate": 4.120392757949954e-08, "loss": 0.972, "step": 5661 }, { "epoch": 3.7923643670462157, "grad_norm": 2.887266938581312, "learning_rate": 4.094010099471968e-08, "loss": 0.9085, "step": 5662 }, { "epoch": 3.793034159410583, "grad_norm": 2.609397320777003, "learning_rate": 4.06771147856988e-08, "loss": 0.7245, "step": 5663 }, { "epoch": 3.79370395177495, "grad_norm": 3.385815565191449, "learning_rate": 4.04149690423114e-08, "loss": 0.8999, "step": 5664 }, { "epoch": 3.7943737441393166, "grad_norm": 3.3000255398462954, "learning_rate": 4.015366385414471e-08, "loss": 0.8935, "step": 5665 }, { "epoch": 3.795043536503684, "grad_norm": 2.950158591699997, "learning_rate": 3.9893199310499785e-08, "loss": 0.9032, "step": 5666 }, { "epoch": 3.795713328868051, "grad_norm": 4.149106757025601, "learning_rate": 3.9633575500388766e-08, "loss": 0.9577, "step": 5667 }, { "epoch": 3.796383121232418, "grad_norm": 3.1061847695832197, "learning_rate": 3.937479251253734e-08, "loss": 0.9225, "step": 5668 }, { "epoch": 3.7970529135967848, "grad_norm": 3.236087325662155, "learning_rate": 3.911685043538393e-08, "loss": 0.8694, "step": 5669 }, { "epoch": 3.797722705961152, "grad_norm": 3.2808578368246146, "learning_rate": 3.885974935707998e-08, "loss": 0.8105, "step": 5670 }, { "epoch": 3.7983924983255193, "grad_norm": 3.082372970869577, "learning_rate": 3.8603489365487966e-08, "loss": 0.7619, "step": 5671 }, { "epoch": 3.799062290689886, "grad_norm": 2.939608173121222, "learning_rate": 3.8348070548184235e-08, "loss": 0.8161, "step": 5672 }, { "epoch": 3.799732083054253, "grad_norm": 3.120742614816842, "learning_rate": 3.8093492992457295e-08, "loss": 0.8087, "step": 5673 }, { "epoch": 3.80040187541862, "grad_norm": 3.2401940601639643, "learning_rate": 3.783975678530866e-08, "loss": 0.7696, "step": 5674 }, { "epoch": 3.8010716677829874, "grad_norm": 3.193750140181807, "learning_rate": 3.7586862013451466e-08, "loss": 0.842, "step": 5675 }, { "epoch": 3.8017414601473543, "grad_norm": 2.934651958832075, "learning_rate": 3.733480876331186e-08, "loss": 0.8579, "step": 5676 }, { "epoch": 3.8024112525117215, "grad_norm": 3.021620259439264, "learning_rate": 3.7083597121027605e-08, "loss": 0.832, "step": 5677 }, { "epoch": 3.8030810448760883, "grad_norm": 2.822778821944542, "learning_rate": 3.683322717245058e-08, "loss": 0.8463, "step": 5678 }, { "epoch": 3.8037508372404556, "grad_norm": 3.3147083927038636, "learning_rate": 3.65836990031429e-08, "loss": 0.804, "step": 5679 }, { "epoch": 3.8044206296048224, "grad_norm": 3.4388309298661808, "learning_rate": 3.633501269838052e-08, "loss": 0.8112, "step": 5680 }, { "epoch": 3.8050904219691897, "grad_norm": 3.2975427611616666, "learning_rate": 3.608716834315101e-08, "loss": 0.9074, "step": 5681 }, { "epoch": 3.8057602143335565, "grad_norm": 2.859635976887555, "learning_rate": 3.584016602215468e-08, "loss": 0.9434, "step": 5682 }, { "epoch": 3.8064300066979238, "grad_norm": 3.4885838863770204, "learning_rate": 3.559400581980316e-08, "loss": 1.0133, "step": 5683 }, { "epoch": 3.8070997990622906, "grad_norm": 2.5733155029602433, "learning_rate": 3.534868782022138e-08, "loss": 0.7165, "step": 5684 }, { "epoch": 3.807769591426658, "grad_norm": 2.92526501582632, "learning_rate": 3.510421210724563e-08, "loss": 0.8995, "step": 5685 }, { "epoch": 3.8084393837910246, "grad_norm": 3.055190827724197, "learning_rate": 3.4860578764425445e-08, "loss": 0.8825, "step": 5686 }, { "epoch": 3.809109176155392, "grad_norm": 3.6850130963739773, "learning_rate": 3.46177878750209e-08, "loss": 0.8852, "step": 5687 }, { "epoch": 3.8097789685197587, "grad_norm": 3.547306070252058, "learning_rate": 3.437583952200535e-08, "loss": 0.8571, "step": 5688 }, { "epoch": 3.810448760884126, "grad_norm": 3.135285192757049, "learning_rate": 3.4134733788063756e-08, "loss": 0.9986, "step": 5689 }, { "epoch": 3.811118553248493, "grad_norm": 2.965943836568379, "learning_rate": 3.389447075559327e-08, "loss": 0.8733, "step": 5690 }, { "epoch": 3.81178834561286, "grad_norm": 2.842211949953674, "learning_rate": 3.3655050506703204e-08, "loss": 0.6483, "step": 5691 }, { "epoch": 3.8124581379772273, "grad_norm": 2.7380170535605406, "learning_rate": 3.341647312321422e-08, "loss": 0.8221, "step": 5692 }, { "epoch": 3.813127930341594, "grad_norm": 2.6806933753602493, "learning_rate": 3.317873868665972e-08, "loss": 0.8689, "step": 5693 }, { "epoch": 3.813797722705961, "grad_norm": 3.437438875876021, "learning_rate": 3.294184727828442e-08, "loss": 0.7659, "step": 5694 }, { "epoch": 3.814467515070328, "grad_norm": 2.986780402934711, "learning_rate": 3.2705798979045524e-08, "loss": 0.8133, "step": 5695 }, { "epoch": 3.8151373074346955, "grad_norm": 2.782719204075917, "learning_rate": 3.247059386961154e-08, "loss": 0.8171, "step": 5696 }, { "epoch": 3.8158070997990623, "grad_norm": 3.3001266104042024, "learning_rate": 3.2236232030362915e-08, "loss": 0.9415, "step": 5697 }, { "epoch": 3.816476892163429, "grad_norm": 2.8412507651236467, "learning_rate": 3.200271354139251e-08, "loss": 0.688, "step": 5698 }, { "epoch": 3.8171466845277964, "grad_norm": 3.014424495561799, "learning_rate": 3.177003848250426e-08, "loss": 1.0179, "step": 5699 }, { "epoch": 3.8178164768921636, "grad_norm": 3.1377448123283513, "learning_rate": 3.1538206933213746e-08, "loss": 0.8513, "step": 5700 }, { "epoch": 3.8184862692565305, "grad_norm": 3.0840354634398475, "learning_rate": 3.130721897274924e-08, "loss": 0.8216, "step": 5701 }, { "epoch": 3.8191560616208973, "grad_norm": 2.98795137773201, "learning_rate": 3.1077074680049836e-08, "loss": 0.9081, "step": 5702 }, { "epoch": 3.8198258539852645, "grad_norm": 2.990450474047171, "learning_rate": 3.084777413376677e-08, "loss": 0.767, "step": 5703 }, { "epoch": 3.820495646349632, "grad_norm": 2.923661287267205, "learning_rate": 3.061931741226237e-08, "loss": 0.7616, "step": 5704 }, { "epoch": 3.8211654387139986, "grad_norm": 2.971710502676413, "learning_rate": 3.0391704593611406e-08, "loss": 1.0215, "step": 5705 }, { "epoch": 3.821835231078366, "grad_norm": 3.0948940356327883, "learning_rate": 3.01649357555997e-08, "loss": 0.8533, "step": 5706 }, { "epoch": 3.8225050234427327, "grad_norm": 3.245372415050295, "learning_rate": 2.993901097572499e-08, "loss": 0.9289, "step": 5707 }, { "epoch": 3.8231748158071, "grad_norm": 3.976187375237972, "learning_rate": 2.971393033119552e-08, "loss": 1.0441, "step": 5708 }, { "epoch": 3.8238446081714668, "grad_norm": 3.9198818966480324, "learning_rate": 2.9489693898932803e-08, "loss": 0.7437, "step": 5709 }, { "epoch": 3.824514400535834, "grad_norm": 2.791141537638479, "learning_rate": 2.9266301755568327e-08, "loss": 0.8772, "step": 5710 }, { "epoch": 3.825184192900201, "grad_norm": 2.9552157327798882, "learning_rate": 2.9043753977446298e-08, "loss": 0.803, "step": 5711 }, { "epoch": 3.825853985264568, "grad_norm": 2.6306074778129687, "learning_rate": 2.88220506406206e-08, "loss": 0.7382, "step": 5712 }, { "epoch": 3.826523777628935, "grad_norm": 3.078619837044114, "learning_rate": 2.8601191820858687e-08, "loss": 0.8497, "step": 5713 }, { "epoch": 3.827193569993302, "grad_norm": 2.9464139790636583, "learning_rate": 2.8381177593637676e-08, "loss": 0.8209, "step": 5714 }, { "epoch": 3.827863362357669, "grad_norm": 2.9299988887269266, "learning_rate": 2.8162008034147425e-08, "loss": 0.9616, "step": 5715 }, { "epoch": 3.8285331547220363, "grad_norm": 2.94720308068942, "learning_rate": 2.7943683217287466e-08, "loss": 0.9356, "step": 5716 }, { "epoch": 3.829202947086403, "grad_norm": 3.132665976551766, "learning_rate": 2.7726203217670333e-08, "loss": 0.8827, "step": 5717 }, { "epoch": 3.8298727394507703, "grad_norm": 3.3009684763482277, "learning_rate": 2.7509568109619077e-08, "loss": 0.8492, "step": 5718 }, { "epoch": 3.830542531815137, "grad_norm": 2.74565375548431, "learning_rate": 2.7293777967167535e-08, "loss": 0.7317, "step": 5719 }, { "epoch": 3.8312123241795044, "grad_norm": 3.592188897959607, "learning_rate": 2.707883286406199e-08, "loss": 0.7226, "step": 5720 }, { "epoch": 3.8318821165438717, "grad_norm": 3.3194040227942327, "learning_rate": 2.6864732873758405e-08, "loss": 0.8915, "step": 5721 }, { "epoch": 3.8325519089082385, "grad_norm": 3.0797876802554924, "learning_rate": 2.6651478069425472e-08, "loss": 0.8488, "step": 5722 }, { "epoch": 3.8332217012726053, "grad_norm": 3.148700419249916, "learning_rate": 2.6439068523942124e-08, "loss": 0.8363, "step": 5723 }, { "epoch": 3.8338914936369726, "grad_norm": 2.87784662596926, "learning_rate": 2.6227504309898345e-08, "loss": 0.773, "step": 5724 }, { "epoch": 3.83456128600134, "grad_norm": 3.759635221464872, "learning_rate": 2.601678549959602e-08, "loss": 0.8418, "step": 5725 }, { "epoch": 3.8352310783657066, "grad_norm": 3.2087029722886182, "learning_rate": 2.580691216504727e-08, "loss": 0.8016, "step": 5726 }, { "epoch": 3.8359008707300735, "grad_norm": 3.0807929587102163, "learning_rate": 2.5597884377975557e-08, "loss": 0.9316, "step": 5727 }, { "epoch": 3.8365706630944407, "grad_norm": 2.9387849069556613, "learning_rate": 2.538970220981568e-08, "loss": 0.7437, "step": 5728 }, { "epoch": 3.837240455458808, "grad_norm": 2.725096678816195, "learning_rate": 2.518236573171323e-08, "loss": 0.7325, "step": 5729 }, { "epoch": 3.837910247823175, "grad_norm": 3.126771870593112, "learning_rate": 2.4975875014524588e-08, "loss": 0.9667, "step": 5730 }, { "epoch": 3.8385800401875416, "grad_norm": 2.9949929740783197, "learning_rate": 2.4770230128817462e-08, "loss": 0.8095, "step": 5731 }, { "epoch": 3.839249832551909, "grad_norm": 3.0169553512113207, "learning_rate": 2.4565431144870368e-08, "loss": 0.8372, "step": 5732 }, { "epoch": 3.839919624916276, "grad_norm": 3.182714072247897, "learning_rate": 2.436147813267231e-08, "loss": 0.8559, "step": 5733 }, { "epoch": 3.840589417280643, "grad_norm": 2.67559619782631, "learning_rate": 2.415837116192421e-08, "loss": 0.812, "step": 5734 }, { "epoch": 3.8412592096450098, "grad_norm": 2.775102670980832, "learning_rate": 2.395611030203665e-08, "loss": 0.811, "step": 5735 }, { "epoch": 3.841929002009377, "grad_norm": 2.948383041253416, "learning_rate": 2.3754695622131564e-08, "loss": 0.8526, "step": 5736 }, { "epoch": 3.8425987943737443, "grad_norm": 3.094326596173709, "learning_rate": 2.35541271910425e-08, "loss": 0.9196, "step": 5737 }, { "epoch": 3.843268586738111, "grad_norm": 4.104396877184054, "learning_rate": 2.3354405077312405e-08, "loss": 0.8512, "step": 5738 }, { "epoch": 3.8439383791024784, "grad_norm": 4.674926788873064, "learning_rate": 2.315552934919585e-08, "loss": 0.8826, "step": 5739 }, { "epoch": 3.844608171466845, "grad_norm": 2.917020744470275, "learning_rate": 2.2957500074657913e-08, "loss": 0.9526, "step": 5740 }, { "epoch": 3.8452779638312125, "grad_norm": 3.1634556247687997, "learning_rate": 2.276031732137446e-08, "loss": 0.7013, "step": 5741 }, { "epoch": 3.8459477561955793, "grad_norm": 3.4420395854296073, "learning_rate": 2.256398115673214e-08, "loss": 0.8131, "step": 5742 }, { "epoch": 3.8466175485599465, "grad_norm": 3.0356813259698843, "learning_rate": 2.2368491647828394e-08, "loss": 0.9427, "step": 5743 }, { "epoch": 3.8472873409243133, "grad_norm": 3.5227418926202434, "learning_rate": 2.2173848861470894e-08, "loss": 0.8093, "step": 5744 }, { "epoch": 3.8479571332886806, "grad_norm": 2.7782845312334246, "learning_rate": 2.198005286417809e-08, "loss": 0.7406, "step": 5745 }, { "epoch": 3.8486269256530474, "grad_norm": 3.157136773379098, "learning_rate": 2.1787103722179237e-08, "loss": 1.0069, "step": 5746 }, { "epoch": 3.8492967180174147, "grad_norm": 3.0862559472170785, "learning_rate": 2.159500150141408e-08, "loss": 0.8158, "step": 5747 }, { "epoch": 3.8499665103817815, "grad_norm": 2.8248411708296546, "learning_rate": 2.1403746267533165e-08, "loss": 0.7263, "step": 5748 }, { "epoch": 3.8506363027461488, "grad_norm": 3.122019194493429, "learning_rate": 2.1213338085896984e-08, "loss": 0.8018, "step": 5749 }, { "epoch": 3.8513060951105156, "grad_norm": 3.1000963765655256, "learning_rate": 2.102377702157682e-08, "loss": 0.7117, "step": 5750 }, { "epoch": 3.851975887474883, "grad_norm": 2.980871051679198, "learning_rate": 2.0835063139355016e-08, "loss": 0.8771, "step": 5751 }, { "epoch": 3.8526456798392497, "grad_norm": 3.0964319629142376, "learning_rate": 2.0647196503723875e-08, "loss": 0.7046, "step": 5752 }, { "epoch": 3.853315472203617, "grad_norm": 2.805909505808167, "learning_rate": 2.0460177178885933e-08, "loss": 0.7474, "step": 5753 }, { "epoch": 3.853985264567984, "grad_norm": 2.916675175569965, "learning_rate": 2.02740052287545e-08, "loss": 0.8377, "step": 5754 }, { "epoch": 3.854655056932351, "grad_norm": 3.2327482510698125, "learning_rate": 2.0088680716953134e-08, "loss": 0.765, "step": 5755 }, { "epoch": 3.855324849296718, "grad_norm": 2.785514478531722, "learning_rate": 1.990420370681645e-08, "loss": 0.8068, "step": 5756 }, { "epoch": 3.855994641661085, "grad_norm": 3.392689675414311, "learning_rate": 1.9720574261388182e-08, "loss": 0.8806, "step": 5757 }, { "epoch": 3.8566644340254523, "grad_norm": 2.6663972991537457, "learning_rate": 1.9537792443423408e-08, "loss": 0.7233, "step": 5758 }, { "epoch": 3.857334226389819, "grad_norm": 3.109697218295956, "learning_rate": 1.9355858315386887e-08, "loss": 0.7912, "step": 5759 }, { "epoch": 3.858004018754186, "grad_norm": 3.0107876968664344, "learning_rate": 1.9174771939454718e-08, "loss": 0.7348, "step": 5760 }, { "epoch": 3.8586738111185532, "grad_norm": 4.309705976261786, "learning_rate": 1.8994533377512393e-08, "loss": 0.6346, "step": 5761 }, { "epoch": 3.8593436034829205, "grad_norm": 3.341081510421018, "learning_rate": 1.8815142691155362e-08, "loss": 0.7748, "step": 5762 }, { "epoch": 3.8600133958472873, "grad_norm": 3.0688405143490876, "learning_rate": 1.863659994168987e-08, "loss": 0.8895, "step": 5763 }, { "epoch": 3.860683188211654, "grad_norm": 2.955790674283501, "learning_rate": 1.8458905190132936e-08, "loss": 0.7843, "step": 5764 }, { "epoch": 3.8613529805760214, "grad_norm": 3.4066847090236796, "learning_rate": 1.828205849721071e-08, "loss": 1.0588, "step": 5765 }, { "epoch": 3.8620227729403886, "grad_norm": 3.6651077230647564, "learning_rate": 1.810605992336012e-08, "loss": 0.9063, "step": 5766 }, { "epoch": 3.8626925653047555, "grad_norm": 3.482922218724254, "learning_rate": 1.793090952872778e-08, "loss": 0.882, "step": 5767 }, { "epoch": 3.8633623576691227, "grad_norm": 3.0574332439127976, "learning_rate": 1.7756607373171087e-08, "loss": 0.9006, "step": 5768 }, { "epoch": 3.8640321500334895, "grad_norm": 2.882233500377129, "learning_rate": 1.7583153516257388e-08, "loss": 0.8615, "step": 5769 }, { "epoch": 3.864701942397857, "grad_norm": 3.071326228826392, "learning_rate": 1.7410548017263717e-08, "loss": 0.9742, "step": 5770 }, { "epoch": 3.8653717347622236, "grad_norm": 2.909431075043205, "learning_rate": 1.723879093517733e-08, "loss": 0.9537, "step": 5771 }, { "epoch": 3.866041527126591, "grad_norm": 3.260336370821907, "learning_rate": 1.7067882328695996e-08, "loss": 0.7806, "step": 5772 }, { "epoch": 3.8667113194909577, "grad_norm": 3.2689098633451246, "learning_rate": 1.689782225622688e-08, "loss": 0.8233, "step": 5773 }, { "epoch": 3.867381111855325, "grad_norm": 2.6838758651468906, "learning_rate": 1.6728610775887667e-08, "loss": 0.7698, "step": 5774 }, { "epoch": 3.8680509042196918, "grad_norm": 2.706664273411323, "learning_rate": 1.6560247945505702e-08, "loss": 0.7804, "step": 5775 }, { "epoch": 3.868720696584059, "grad_norm": 2.7482440207179137, "learning_rate": 1.639273382261858e-08, "loss": 0.8041, "step": 5776 }, { "epoch": 3.869390488948426, "grad_norm": 4.722417626381678, "learning_rate": 1.6226068464473555e-08, "loss": 0.8434, "step": 5777 }, { "epoch": 3.870060281312793, "grad_norm": 3.132720391137255, "learning_rate": 1.6060251928028127e-08, "loss": 0.8493, "step": 5778 }, { "epoch": 3.87073007367716, "grad_norm": 2.8761387441433235, "learning_rate": 1.5895284269949184e-08, "loss": 0.8173, "step": 5779 }, { "epoch": 3.871399866041527, "grad_norm": 2.9181731934278035, "learning_rate": 1.5731165546614412e-08, "loss": 0.9604, "step": 5780 }, { "epoch": 3.872069658405894, "grad_norm": 3.0642773806227517, "learning_rate": 1.5567895814110888e-08, "loss": 1.0143, "step": 5781 }, { "epoch": 3.8727394507702613, "grad_norm": 2.8751555611967365, "learning_rate": 1.5405475128234813e-08, "loss": 0.8025, "step": 5782 }, { "epoch": 3.8734092431346285, "grad_norm": 4.366023042236898, "learning_rate": 1.5243903544493454e-08, "loss": 0.9093, "step": 5783 }, { "epoch": 3.8740790354989953, "grad_norm": 3.0492694357059773, "learning_rate": 1.5083181118103195e-08, "loss": 0.9455, "step": 5784 }, { "epoch": 3.874748827863362, "grad_norm": 2.975214374235729, "learning_rate": 1.4923307903990936e-08, "loss": 0.8223, "step": 5785 }, { "epoch": 3.8754186202277294, "grad_norm": 2.9928705985672184, "learning_rate": 1.4764283956791859e-08, "loss": 0.8486, "step": 5786 }, { "epoch": 3.8760884125920967, "grad_norm": 3.2077418806528906, "learning_rate": 1.4606109330852492e-08, "loss": 0.8756, "step": 5787 }, { "epoch": 3.8767582049564635, "grad_norm": 3.050006909268422, "learning_rate": 1.4448784080228762e-08, "loss": 0.8292, "step": 5788 }, { "epoch": 3.8774279973208303, "grad_norm": 2.6424285413586164, "learning_rate": 1.4292308258685439e-08, "loss": 0.8374, "step": 5789 }, { "epoch": 3.8780977896851976, "grad_norm": 4.141144844786236, "learning_rate": 1.4136681919698081e-08, "loss": 0.8793, "step": 5790 }, { "epoch": 3.878767582049565, "grad_norm": 3.3646390295000788, "learning_rate": 1.3981905116451089e-08, "loss": 0.938, "step": 5791 }, { "epoch": 3.8794373744139317, "grad_norm": 2.6275478477614502, "learning_rate": 1.3827977901839373e-08, "loss": 0.8135, "step": 5792 }, { "epoch": 3.8801071667782985, "grad_norm": 2.9216874379240205, "learning_rate": 1.367490032846669e-08, "loss": 0.8001, "step": 5793 }, { "epoch": 3.8807769591426657, "grad_norm": 3.458337821788972, "learning_rate": 1.3522672448647022e-08, "loss": 0.9862, "step": 5794 }, { "epoch": 3.881446751507033, "grad_norm": 3.03931716520956, "learning_rate": 1.337129431440376e-08, "loss": 0.9772, "step": 5795 }, { "epoch": 3.8821165438714, "grad_norm": 3.0241119918882955, "learning_rate": 1.3220765977469686e-08, "loss": 0.8009, "step": 5796 }, { "epoch": 3.882786336235767, "grad_norm": 3.2922928514506444, "learning_rate": 1.3071087489287814e-08, "loss": 0.931, "step": 5797 }, { "epoch": 3.883456128600134, "grad_norm": 3.6474504075216694, "learning_rate": 1.2922258901009732e-08, "loss": 0.8282, "step": 5798 }, { "epoch": 3.884125920964501, "grad_norm": 3.2386398195771395, "learning_rate": 1.2774280263497252e-08, "loss": 0.8373, "step": 5799 }, { "epoch": 3.884795713328868, "grad_norm": 2.834512385824001, "learning_rate": 1.2627151627322143e-08, "loss": 0.9046, "step": 5800 }, { "epoch": 3.8854655056932352, "grad_norm": 2.928110577078105, "learning_rate": 1.2480873042764464e-08, "loss": 0.7668, "step": 5801 }, { "epoch": 3.886135298057602, "grad_norm": 3.0452421701574077, "learning_rate": 1.2335444559814502e-08, "loss": 0.9586, "step": 5802 }, { "epoch": 3.8868050904219693, "grad_norm": 3.2611094147527893, "learning_rate": 1.2190866228172505e-08, "loss": 0.9098, "step": 5803 }, { "epoch": 3.887474882786336, "grad_norm": 3.12329052012588, "learning_rate": 1.204713809724728e-08, "loss": 0.9341, "step": 5804 }, { "epoch": 3.8881446751507034, "grad_norm": 2.671943767820482, "learning_rate": 1.1904260216157314e-08, "loss": 0.8619, "step": 5805 }, { "epoch": 3.88881446751507, "grad_norm": 2.9966373139135873, "learning_rate": 1.1762232633730775e-08, "loss": 0.9023, "step": 5806 }, { "epoch": 3.8894842598794375, "grad_norm": 3.3628688443220196, "learning_rate": 1.1621055398505498e-08, "loss": 0.9113, "step": 5807 }, { "epoch": 3.8901540522438043, "grad_norm": 3.5214274844514475, "learning_rate": 1.1480728558727894e-08, "loss": 0.948, "step": 5808 }, { "epoch": 3.8908238446081715, "grad_norm": 3.746596775580209, "learning_rate": 1.1341252162354322e-08, "loss": 0.7953, "step": 5809 }, { "epoch": 3.8914936369725384, "grad_norm": 3.2344031662918096, "learning_rate": 1.1202626257050264e-08, "loss": 1.0169, "step": 5810 }, { "epoch": 3.8921634293369056, "grad_norm": 2.8745046296808674, "learning_rate": 1.1064850890190881e-08, "loss": 0.8942, "step": 5811 }, { "epoch": 3.8928332217012724, "grad_norm": 3.0258642369374797, "learning_rate": 1.0927926108860454e-08, "loss": 0.798, "step": 5812 }, { "epoch": 3.8935030140656397, "grad_norm": 3.0245082991580627, "learning_rate": 1.0791851959852662e-08, "loss": 0.7679, "step": 5813 }, { "epoch": 3.8941728064300065, "grad_norm": 2.9689500270875127, "learning_rate": 1.0656628489670307e-08, "loss": 0.9741, "step": 5814 }, { "epoch": 3.894842598794374, "grad_norm": 2.6824763812267682, "learning_rate": 1.0522255744525311e-08, "loss": 0.7863, "step": 5815 }, { "epoch": 3.895512391158741, "grad_norm": 3.1663533870478298, "learning_rate": 1.0388733770339276e-08, "loss": 0.8842, "step": 5816 }, { "epoch": 3.896182183523108, "grad_norm": 3.525920293786613, "learning_rate": 1.02560626127432e-08, "loss": 1.0166, "step": 5817 }, { "epoch": 3.8968519758874747, "grad_norm": 3.197469768438637, "learning_rate": 1.0124242317076927e-08, "loss": 0.9608, "step": 5818 }, { "epoch": 3.897521768251842, "grad_norm": 3.4143857166298957, "learning_rate": 9.993272928389141e-09, "loss": 0.9406, "step": 5819 }, { "epoch": 3.898191560616209, "grad_norm": 2.8821596439549624, "learning_rate": 9.863154491438765e-09, "loss": 0.8326, "step": 5820 }, { "epoch": 3.898861352980576, "grad_norm": 2.9076926388317936, "learning_rate": 9.733887050693003e-09, "loss": 0.8055, "step": 5821 }, { "epoch": 3.899531145344943, "grad_norm": 2.918645950339401, "learning_rate": 9.605470650329019e-09, "loss": 0.8596, "step": 5822 }, { "epoch": 3.90020093770931, "grad_norm": 2.94622162518232, "learning_rate": 9.477905334232263e-09, "loss": 0.8875, "step": 5823 }, { "epoch": 3.9008707300736774, "grad_norm": 2.9423265741975064, "learning_rate": 9.351191145997861e-09, "loss": 0.8107, "step": 5824 }, { "epoch": 3.901540522438044, "grad_norm": 3.296292087876658, "learning_rate": 9.225328128930066e-09, "loss": 0.9051, "step": 5825 }, { "epoch": 3.902210314802411, "grad_norm": 3.1214412768889326, "learning_rate": 9.100316326041969e-09, "loss": 0.8563, "step": 5826 }, { "epoch": 3.9028801071667782, "grad_norm": 2.8061780068697972, "learning_rate": 8.97615578005634e-09, "loss": 0.8457, "step": 5827 }, { "epoch": 3.9035498995311455, "grad_norm": 2.8447027285522157, "learning_rate": 8.852846533404235e-09, "loss": 0.7484, "step": 5828 }, { "epoch": 3.9042196918955123, "grad_norm": 2.796703046558478, "learning_rate": 8.730388628226394e-09, "loss": 0.8137, "step": 5829 }, { "epoch": 3.9048894842598796, "grad_norm": 4.2663308290560344, "learning_rate": 8.608782106372394e-09, "loss": 0.8175, "step": 5830 }, { "epoch": 3.9055592766242464, "grad_norm": 3.5396238478220203, "learning_rate": 8.488027009400934e-09, "loss": 0.9777, "step": 5831 }, { "epoch": 3.9062290689886137, "grad_norm": 2.5606484621013412, "learning_rate": 8.368123378579285e-09, "loss": 0.703, "step": 5832 }, { "epoch": 3.9068988613529805, "grad_norm": 3.5729551082661755, "learning_rate": 8.249071254884666e-09, "loss": 0.9558, "step": 5833 }, { "epoch": 3.9075686537173477, "grad_norm": 3.626339877661685, "learning_rate": 8.13087067900259e-09, "loss": 0.8267, "step": 5834 }, { "epoch": 3.9082384460817146, "grad_norm": 2.6663359989309026, "learning_rate": 8.013521691327686e-09, "loss": 0.7913, "step": 5835 }, { "epoch": 3.908908238446082, "grad_norm": 3.060268218691827, "learning_rate": 7.897024331963432e-09, "loss": 0.8092, "step": 5836 }, { "epoch": 3.9095780308104486, "grad_norm": 3.405571478430217, "learning_rate": 7.781378640722704e-09, "loss": 0.9506, "step": 5837 }, { "epoch": 3.910247823174816, "grad_norm": 3.0226326115928916, "learning_rate": 7.666584657126663e-09, "loss": 0.8518, "step": 5838 }, { "epoch": 3.9109176155391827, "grad_norm": 3.1442168466767764, "learning_rate": 7.552642420406154e-09, "loss": 0.9846, "step": 5839 }, { "epoch": 3.91158740790355, "grad_norm": 3.4612959270163564, "learning_rate": 7.4395519695003026e-09, "loss": 0.884, "step": 5840 }, { "epoch": 3.912257200267917, "grad_norm": 3.1041309925963403, "learning_rate": 7.32731334305764e-09, "loss": 0.8838, "step": 5841 }, { "epoch": 3.912926992632284, "grad_norm": 3.1739905383435305, "learning_rate": 7.215926579434984e-09, "loss": 1.0456, "step": 5842 }, { "epoch": 3.913596784996651, "grad_norm": 3.037381761032206, "learning_rate": 7.105391716699106e-09, "loss": 0.7286, "step": 5843 }, { "epoch": 3.914266577361018, "grad_norm": 2.716287749918404, "learning_rate": 6.995708792624234e-09, "loss": 0.7705, "step": 5844 }, { "epoch": 3.9149363697253854, "grad_norm": 2.8278353867338706, "learning_rate": 6.886877844694273e-09, "loss": 0.9246, "step": 5845 }, { "epoch": 3.915606162089752, "grad_norm": 3.8693066536872243, "learning_rate": 6.778898910102249e-09, "loss": 0.8348, "step": 5846 }, { "epoch": 3.916275954454119, "grad_norm": 3.417763339979475, "learning_rate": 6.671772025749201e-09, "loss": 1.0695, "step": 5847 }, { "epoch": 3.9169457468184863, "grad_norm": 3.012623210174206, "learning_rate": 6.565497228245565e-09, "loss": 0.7727, "step": 5848 }, { "epoch": 3.9176155391828535, "grad_norm": 3.0879872340203525, "learning_rate": 6.460074553910067e-09, "loss": 0.8786, "step": 5849 }, { "epoch": 3.9182853315472204, "grad_norm": 2.7082586449710475, "learning_rate": 6.355504038771387e-09, "loss": 0.752, "step": 5850 }, { "epoch": 3.918955123911587, "grad_norm": 2.5375901986569236, "learning_rate": 6.251785718565384e-09, "loss": 0.7268, "step": 5851 }, { "epoch": 3.9196249162759544, "grad_norm": 3.1438160886178292, "learning_rate": 6.14891962873787e-09, "loss": 0.824, "step": 5852 }, { "epoch": 3.9202947086403217, "grad_norm": 2.723948162355126, "learning_rate": 6.046905804442671e-09, "loss": 0.8613, "step": 5853 }, { "epoch": 3.9209645010046885, "grad_norm": 2.7182396792826675, "learning_rate": 5.945744280542731e-09, "loss": 0.8775, "step": 5854 }, { "epoch": 3.9216342933690553, "grad_norm": 11.60142873681629, "learning_rate": 5.845435091610119e-09, "loss": 0.8014, "step": 5855 }, { "epoch": 3.9223040857334226, "grad_norm": 3.0998931207618905, "learning_rate": 5.745978271924912e-09, "loss": 0.7677, "step": 5856 }, { "epoch": 3.92297387809779, "grad_norm": 4.212557425376009, "learning_rate": 5.647373855475757e-09, "loss": 0.8795, "step": 5857 }, { "epoch": 3.9236436704621567, "grad_norm": 2.6309825704588152, "learning_rate": 5.549621875960975e-09, "loss": 0.8091, "step": 5858 }, { "epoch": 3.924313462826524, "grad_norm": 2.6421062688774635, "learning_rate": 5.452722366786622e-09, "loss": 0.7566, "step": 5859 }, { "epoch": 3.9249832551908908, "grad_norm": 2.783891980941311, "learning_rate": 5.356675361068153e-09, "loss": 0.7772, "step": 5860 }, { "epoch": 3.925653047555258, "grad_norm": 2.9672802649555945, "learning_rate": 5.261480891628756e-09, "loss": 0.8775, "step": 5861 }, { "epoch": 3.926322839919625, "grad_norm": 2.991758218117331, "learning_rate": 5.167138991001297e-09, "loss": 0.8462, "step": 5862 }, { "epoch": 3.926992632283992, "grad_norm": 3.0445715381072924, "learning_rate": 5.073649691426929e-09, "loss": 0.8207, "step": 5863 }, { "epoch": 3.927662424648359, "grad_norm": 2.9810031437651134, "learning_rate": 4.9810130248548195e-09, "loss": 0.8634, "step": 5864 }, { "epoch": 3.928332217012726, "grad_norm": 3.130860052304187, "learning_rate": 4.889229022943531e-09, "loss": 0.9991, "step": 5865 }, { "epoch": 3.929002009377093, "grad_norm": 2.9419099595279055, "learning_rate": 4.7982977170599185e-09, "loss": 0.7417, "step": 5866 }, { "epoch": 3.9296718017414602, "grad_norm": 2.9455629584317418, "learning_rate": 4.708219138279679e-09, "loss": 0.8711, "step": 5867 }, { "epoch": 3.930341594105827, "grad_norm": 3.4759547573162695, "learning_rate": 4.618993317386522e-09, "loss": 1.0123, "step": 5868 }, { "epoch": 3.9310113864701943, "grad_norm": 3.3193632635658186, "learning_rate": 4.530620284873e-09, "loss": 0.8885, "step": 5869 }, { "epoch": 3.931681178834561, "grad_norm": 3.4565674843235046, "learning_rate": 4.443100070940787e-09, "loss": 0.8894, "step": 5870 }, { "epoch": 3.9323509711989284, "grad_norm": 3.2318498945879734, "learning_rate": 4.356432705499292e-09, "loss": 0.9772, "step": 5871 }, { "epoch": 3.933020763563295, "grad_norm": 2.79477591687838, "learning_rate": 4.270618218167044e-09, "loss": 0.7821, "step": 5872 }, { "epoch": 3.9336905559276625, "grad_norm": 4.370174502622064, "learning_rate": 4.1856566382705855e-09, "loss": 0.7832, "step": 5873 }, { "epoch": 3.9343603482920297, "grad_norm": 3.893972155411634, "learning_rate": 4.101547994845579e-09, "loss": 0.8422, "step": 5874 }, { "epoch": 3.9350301406563966, "grad_norm": 2.5961572061650573, "learning_rate": 4.018292316635419e-09, "loss": 0.71, "step": 5875 }, { "epoch": 3.9356999330207634, "grad_norm": 2.898167085482802, "learning_rate": 3.935889632092904e-09, "loss": 0.7536, "step": 5876 }, { "epoch": 3.9363697253851306, "grad_norm": 2.710521887705358, "learning_rate": 3.854339969378562e-09, "loss": 0.796, "step": 5877 }, { "epoch": 3.937039517749498, "grad_norm": 6.121922249393996, "learning_rate": 3.773643356362045e-09, "loss": 0.9769, "step": 5878 }, { "epoch": 3.9377093101138647, "grad_norm": 2.8581927935125573, "learning_rate": 3.6937998206210135e-09, "loss": 0.8446, "step": 5879 }, { "epoch": 3.9383791024782315, "grad_norm": 3.9041876422840534, "learning_rate": 3.614809389441698e-09, "loss": 0.7186, "step": 5880 }, { "epoch": 3.939048894842599, "grad_norm": 3.0249086263199594, "learning_rate": 3.5366720898186158e-09, "loss": 0.8315, "step": 5881 }, { "epoch": 3.939718687206966, "grad_norm": 2.9786182202363896, "learning_rate": 3.4593879484554062e-09, "loss": 0.8659, "step": 5882 }, { "epoch": 3.940388479571333, "grad_norm": 2.8984434092635025, "learning_rate": 3.382956991763442e-09, "loss": 0.8701, "step": 5883 }, { "epoch": 3.9410582719356997, "grad_norm": 2.7674100027447612, "learning_rate": 3.307379245862663e-09, "loss": 0.8147, "step": 5884 }, { "epoch": 3.941728064300067, "grad_norm": 2.660627702218222, "learning_rate": 3.2326547365815752e-09, "loss": 0.8211, "step": 5885 }, { "epoch": 3.942397856664434, "grad_norm": 3.168859180769569, "learning_rate": 3.158783489457251e-09, "loss": 0.8393, "step": 5886 }, { "epoch": 3.943067649028801, "grad_norm": 4.617648506668249, "learning_rate": 3.085765529734497e-09, "loss": 0.8087, "step": 5887 }, { "epoch": 3.943737441393168, "grad_norm": 2.950666783202327, "learning_rate": 3.0136008823675177e-09, "loss": 0.86, "step": 5888 }, { "epoch": 3.944407233757535, "grad_norm": 3.242633169268622, "learning_rate": 2.9422895720179755e-09, "loss": 0.7287, "step": 5889 }, { "epoch": 3.9450770261219024, "grad_norm": 3.3143740856337334, "learning_rate": 2.8718316230560984e-09, "loss": 0.775, "step": 5890 }, { "epoch": 3.945746818486269, "grad_norm": 2.860306561362046, "learning_rate": 2.8022270595612356e-09, "loss": 0.9111, "step": 5891 }, { "epoch": 3.9464166108506364, "grad_norm": 3.245607417505759, "learning_rate": 2.733475905320193e-09, "loss": 0.8488, "step": 5892 }, { "epoch": 3.9470864032150033, "grad_norm": 3.261081676414159, "learning_rate": 2.6655781838283436e-09, "loss": 0.8446, "step": 5893 }, { "epoch": 3.9477561955793705, "grad_norm": 4.162217558179933, "learning_rate": 2.5985339182896252e-09, "loss": 0.9691, "step": 5894 }, { "epoch": 3.9484259879437373, "grad_norm": 3.068079117977817, "learning_rate": 2.532343131616544e-09, "loss": 0.8631, "step": 5895 }, { "epoch": 3.9490957803081046, "grad_norm": 3.2403336160071445, "learning_rate": 2.46700584642906e-09, "loss": 0.8419, "step": 5896 }, { "epoch": 3.9497655726724714, "grad_norm": 2.724421067130134, "learning_rate": 2.4025220850559805e-09, "loss": 0.7756, "step": 5897 }, { "epoch": 3.9504353650368387, "grad_norm": 3.1461965777111254, "learning_rate": 2.338891869534954e-09, "loss": 0.8334, "step": 5898 }, { "epoch": 3.9511051574012055, "grad_norm": 2.8463702141647915, "learning_rate": 2.2761152216108085e-09, "loss": 0.8468, "step": 5899 }, { "epoch": 3.9517749497655728, "grad_norm": 3.1525778067834374, "learning_rate": 2.2141921627377717e-09, "loss": 0.7722, "step": 5900 }, { "epoch": 3.9524447421299396, "grad_norm": 3.3570778369970853, "learning_rate": 2.153122714077527e-09, "loss": 1.029, "step": 5901 }, { "epoch": 3.953114534494307, "grad_norm": 2.855103935495849, "learning_rate": 2.092906896500324e-09, "loss": 0.8607, "step": 5902 }, { "epoch": 3.9537843268586736, "grad_norm": 3.6560055781717797, "learning_rate": 2.03354473058498e-09, "loss": 0.9218, "step": 5903 }, { "epoch": 3.954454119223041, "grad_norm": 3.110219865332811, "learning_rate": 1.9750362366180444e-09, "loss": 0.856, "step": 5904 }, { "epoch": 3.9551239115874077, "grad_norm": 2.9279203859485587, "learning_rate": 1.9173814345949116e-09, "loss": 0.8694, "step": 5905 }, { "epoch": 3.955793703951775, "grad_norm": 2.810923108701708, "learning_rate": 1.8605803442184322e-09, "loss": 0.8537, "step": 5906 }, { "epoch": 3.9564634963161422, "grad_norm": 3.0312907349345743, "learning_rate": 1.804632984900856e-09, "loss": 0.8535, "step": 5907 }, { "epoch": 3.957133288680509, "grad_norm": 2.5830948241348186, "learning_rate": 1.7495393757616108e-09, "loss": 0.8286, "step": 5908 }, { "epoch": 3.957803081044876, "grad_norm": 3.148484624623945, "learning_rate": 1.6952995356286916e-09, "loss": 0.9301, "step": 5909 }, { "epoch": 3.958472873409243, "grad_norm": 2.8244526058337973, "learning_rate": 1.6419134830386595e-09, "loss": 0.6989, "step": 5910 }, { "epoch": 3.9591426657736104, "grad_norm": 3.183951843597871, "learning_rate": 1.5893812362355321e-09, "loss": 0.8298, "step": 5911 }, { "epoch": 3.959812458137977, "grad_norm": 3.256657299921077, "learning_rate": 1.5377028131727256e-09, "loss": 0.6929, "step": 5912 }, { "epoch": 3.960482250502344, "grad_norm": 3.2499399441064916, "learning_rate": 1.4868782315108354e-09, "loss": 0.8734, "step": 5913 }, { "epoch": 3.9611520428667113, "grad_norm": 3.2281613882342555, "learning_rate": 1.4369075086187457e-09, "loss": 0.8989, "step": 5914 }, { "epoch": 3.9618218352310786, "grad_norm": 3.337644434440924, "learning_rate": 1.3877906615741843e-09, "loss": 0.9209, "step": 5915 }, { "epoch": 3.9624916275954454, "grad_norm": 2.827352521344691, "learning_rate": 1.3395277071626133e-09, "loss": 0.8277, "step": 5916 }, { "epoch": 3.963161419959812, "grad_norm": 3.4005944004216944, "learning_rate": 1.2921186618777836e-09, "loss": 0.9689, "step": 5917 }, { "epoch": 3.9638312123241795, "grad_norm": 2.986733621516334, "learning_rate": 1.2455635419211797e-09, "loss": 0.8508, "step": 5918 }, { "epoch": 3.9645010046885467, "grad_norm": 3.0843227509704776, "learning_rate": 1.1998623632031304e-09, "loss": 0.8844, "step": 5919 }, { "epoch": 3.9651707970529135, "grad_norm": 2.629873899366037, "learning_rate": 1.1550151413419753e-09, "loss": 0.7071, "step": 5920 }, { "epoch": 3.965840589417281, "grad_norm": 2.7793823208269086, "learning_rate": 1.1110218916640658e-09, "loss": 0.7973, "step": 5921 }, { "epoch": 3.9665103817816476, "grad_norm": 2.6295908625196995, "learning_rate": 1.0678826292037648e-09, "loss": 0.8991, "step": 5922 }, { "epoch": 3.967180174146015, "grad_norm": 2.778242869521229, "learning_rate": 1.025597368704001e-09, "loss": 0.6835, "step": 5923 }, { "epoch": 3.9678499665103817, "grad_norm": 3.326586526551058, "learning_rate": 9.841661246154377e-10, "loss": 0.8991, "step": 5924 }, { "epoch": 3.968519758874749, "grad_norm": 2.899658733609495, "learning_rate": 9.435889110970265e-10, "loss": 0.7652, "step": 5925 }, { "epoch": 3.9691895512391158, "grad_norm": 2.795011127908357, "learning_rate": 9.038657420162855e-10, "loss": 0.9041, "step": 5926 }, { "epoch": 3.969859343603483, "grad_norm": 2.952550827972206, "learning_rate": 8.649966309481894e-10, "loss": 0.6753, "step": 5927 }, { "epoch": 3.97052913596785, "grad_norm": 3.0563856214017084, "learning_rate": 8.269815911760016e-10, "loss": 0.8095, "step": 5928 }, { "epoch": 3.971198928332217, "grad_norm": 2.7620812150168175, "learning_rate": 7.898206356915516e-10, "loss": 0.7716, "step": 5929 }, { "epoch": 3.971868720696584, "grad_norm": 2.818247374120143, "learning_rate": 7.535137771944034e-10, "loss": 0.7502, "step": 5930 }, { "epoch": 3.972538513060951, "grad_norm": 2.995769305069727, "learning_rate": 7.180610280921318e-10, "loss": 0.832, "step": 5931 }, { "epoch": 3.973208305425318, "grad_norm": 2.871727790622584, "learning_rate": 6.834624005006008e-10, "loss": 0.8324, "step": 5932 }, { "epoch": 3.9738780977896853, "grad_norm": 2.906952569350712, "learning_rate": 6.497179062439629e-10, "loss": 0.9278, "step": 5933 }, { "epoch": 3.974547890154052, "grad_norm": 3.1216231983655662, "learning_rate": 6.168275568541049e-10, "loss": 0.8193, "step": 5934 }, { "epoch": 3.9752176825184193, "grad_norm": 2.6244369631761875, "learning_rate": 5.847913635714797e-10, "loss": 0.7126, "step": 5935 }, { "epoch": 3.9758874748827866, "grad_norm": 3.22348615920429, "learning_rate": 5.536093373439965e-10, "loss": 0.9505, "step": 5936 }, { "epoch": 3.9765572672471534, "grad_norm": 3.0032307100714086, "learning_rate": 5.232814888284088e-10, "loss": 0.8632, "step": 5937 }, { "epoch": 3.9772270596115202, "grad_norm": 3.8681830557017185, "learning_rate": 4.938078283889258e-10, "loss": 0.8826, "step": 5938 }, { "epoch": 3.9778968519758875, "grad_norm": 3.2472817404729395, "learning_rate": 4.6518836609776853e-10, "loss": 0.8437, "step": 5939 }, { "epoch": 3.9785666443402548, "grad_norm": 3.1200458905876984, "learning_rate": 4.374231117360017e-10, "loss": 0.7848, "step": 5940 }, { "epoch": 3.9792364367046216, "grad_norm": 2.992454958899536, "learning_rate": 4.1051207479242404e-10, "loss": 0.8984, "step": 5941 }, { "epoch": 3.9799062290689884, "grad_norm": 2.864243523219895, "learning_rate": 3.844552644632904e-10, "loss": 0.7953, "step": 5942 }, { "epoch": 3.9805760214333556, "grad_norm": 2.9497548095938617, "learning_rate": 3.5925268965369966e-10, "loss": 0.7696, "step": 5943 }, { "epoch": 3.981245813797723, "grad_norm": 2.7650491653484255, "learning_rate": 3.3490435897648444e-10, "loss": 0.6714, "step": 5944 }, { "epoch": 3.9819156061620897, "grad_norm": 2.8578494293077297, "learning_rate": 3.114102807527664e-10, "loss": 0.9467, "step": 5945 }, { "epoch": 3.9825853985264565, "grad_norm": 3.6335231783064033, "learning_rate": 2.8877046301167833e-10, "loss": 0.894, "step": 5946 }, { "epoch": 3.983255190890824, "grad_norm": 4.038456932946974, "learning_rate": 2.669849134898095e-10, "loss": 0.9142, "step": 5947 }, { "epoch": 3.983924983255191, "grad_norm": 2.9298781136346324, "learning_rate": 2.4605363963259297e-10, "loss": 0.938, "step": 5948 }, { "epoch": 3.984594775619558, "grad_norm": 2.6886383279408723, "learning_rate": 2.2597664859319578e-10, "loss": 0.8027, "step": 5949 }, { "epoch": 3.985264567983925, "grad_norm": 3.325340779132509, "learning_rate": 2.0675394723307374e-10, "loss": 0.9221, "step": 5950 }, { "epoch": 3.985934360348292, "grad_norm": 3.035598253185242, "learning_rate": 1.883855421211389e-10, "loss": 0.9267, "step": 5951 }, { "epoch": 3.986604152712659, "grad_norm": 3.0263573928005796, "learning_rate": 1.7087143953486983e-10, "loss": 0.7518, "step": 5952 }, { "epoch": 3.987273945077026, "grad_norm": 3.1344271167578612, "learning_rate": 1.542116454597564e-10, "loss": 0.852, "step": 5953 }, { "epoch": 3.9879437374413933, "grad_norm": 2.664667079872829, "learning_rate": 1.3840616558929988e-10, "loss": 0.7067, "step": 5954 }, { "epoch": 3.98861352980576, "grad_norm": 3.096467886358314, "learning_rate": 1.234550053247352e-10, "loss": 0.8584, "step": 5955 }, { "epoch": 3.9892833221701274, "grad_norm": 2.6915511458014314, "learning_rate": 1.0935816977586389e-10, "loss": 0.7721, "step": 5956 }, { "epoch": 3.989953114534494, "grad_norm": 3.5095196077281936, "learning_rate": 9.611566375994363e-11, "loss": 0.8232, "step": 5957 }, { "epoch": 3.9906229068988615, "grad_norm": 2.9673721177693575, "learning_rate": 8.37274918027986e-11, "loss": 0.9316, "step": 5958 }, { "epoch": 3.9912926992632283, "grad_norm": 3.2915998831934417, "learning_rate": 7.219365813770917e-11, "loss": 0.9033, "step": 5959 }, { "epoch": 3.9919624916275955, "grad_norm": 3.3596297298364934, "learning_rate": 6.151416670679977e-11, "loss": 0.9169, "step": 5960 }, { "epoch": 3.9926322839919623, "grad_norm": 3.2185598011016054, "learning_rate": 5.168902115909591e-11, "loss": 0.9255, "step": 5961 }, { "epoch": 3.9933020763563296, "grad_norm": 3.143933620722814, "learning_rate": 4.271822485302224e-11, "loss": 0.9259, "step": 5962 }, { "epoch": 3.9939718687206964, "grad_norm": 2.9459689475407296, "learning_rate": 3.460178085390453e-11, "loss": 0.827, "step": 5963 }, { "epoch": 3.9946416610850637, "grad_norm": 3.3941517082617527, "learning_rate": 2.7339691935635014e-11, "loss": 0.9932, "step": 5964 }, { "epoch": 3.995311453449431, "grad_norm": 3.1086098826986457, "learning_rate": 2.093196058011726e-11, "loss": 0.8445, "step": 5965 }, { "epoch": 3.9959812458137978, "grad_norm": 3.2615151016632713, "learning_rate": 1.537858897698863e-11, "loss": 0.9064, "step": 5966 }, { "epoch": 3.9966510381781646, "grad_norm": 3.028304269676416, "learning_rate": 1.0679579023897823e-11, "loss": 0.7623, "step": 5967 }, { "epoch": 3.997320830542532, "grad_norm": 2.9473359671699844, "learning_rate": 6.8349323270600064e-12, "loss": 0.8068, "step": 5968 }, { "epoch": 3.997990622906899, "grad_norm": 3.1266430521532222, "learning_rate": 3.844650200424127e-12, "loss": 0.9485, "step": 5969 }, { "epoch": 3.998660415271266, "grad_norm": 3.1701638341907885, "learning_rate": 1.7087336656729237e-12, "loss": 0.8547, "step": 5970 }, { "epoch": 3.9993302076356327, "grad_norm": 3.0111780692173564, "learning_rate": 4.271834530555907e-13, "loss": 0.9418, "step": 5971 }, { "epoch": 4.0, "grad_norm": 2.6670515782339765, "learning_rate": 0.0, "loss": 0.7042, "step": 5972 } ], "logging_steps": 1.0, "max_steps": 5972, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 564135589011456.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }