{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9983633387888707, "eval_steps": 500, "global_step": 305, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0032733224222585926, "grad_norm": 4.5458229969104, "learning_rate": 3.3333333333333335e-07, "loss": 2.5273, "step": 1 }, { "epoch": 0.006546644844517185, "grad_norm": 4.12339411074909, "learning_rate": 6.666666666666667e-07, "loss": 2.5212, "step": 2 }, { "epoch": 0.009819967266775777, "grad_norm": 7.698312038734646, "learning_rate": 1.0000000000000002e-06, "loss": 2.553, "step": 3 }, { "epoch": 0.01309328968903437, "grad_norm": 3.2743265686425325, "learning_rate": 1.3333333333333334e-06, "loss": 2.5286, "step": 4 }, { "epoch": 0.016366612111292964, "grad_norm": 4.882632956145234, "learning_rate": 1.6666666666666667e-06, "loss": 2.5099, "step": 5 }, { "epoch": 0.019639934533551555, "grad_norm": 1.5422971469796205, "learning_rate": 2.0000000000000003e-06, "loss": 2.5156, "step": 6 }, { "epoch": 0.022913256955810146, "grad_norm": 4.3292940294795, "learning_rate": 2.3333333333333336e-06, "loss": 2.5432, "step": 7 }, { "epoch": 0.02618657937806874, "grad_norm": 1.3106542988714809, "learning_rate": 2.666666666666667e-06, "loss": 2.5001, "step": 8 }, { "epoch": 0.029459901800327332, "grad_norm": 1.915309733377902, "learning_rate": 3e-06, "loss": 2.5289, "step": 9 }, { "epoch": 0.03273322422258593, "grad_norm": 1.4811295343620543, "learning_rate": 3.3333333333333333e-06, "loss": 2.5249, "step": 10 }, { "epoch": 0.03600654664484452, "grad_norm": 0.9585352491709347, "learning_rate": 3.6666666666666666e-06, "loss": 2.5195, "step": 11 }, { "epoch": 0.03927986906710311, "grad_norm": 0.6542974173213092, "learning_rate": 4.000000000000001e-06, "loss": 2.5096, "step": 12 }, { "epoch": 0.0425531914893617, "grad_norm": 0.6159788634871752, "learning_rate": 4.333333333333334e-06, "loss": 2.509, "step": 13 }, { "epoch": 0.04582651391162029, "grad_norm": 0.6302127259561674, "learning_rate": 4.666666666666667e-06, "loss": 2.5159, "step": 14 }, { "epoch": 0.049099836333878884, "grad_norm": 0.5116118125911384, "learning_rate": 5e-06, "loss": 2.4919, "step": 15 }, { "epoch": 0.05237315875613748, "grad_norm": 0.42670850345300865, "learning_rate": 5.333333333333334e-06, "loss": 2.5056, "step": 16 }, { "epoch": 0.05564648117839607, "grad_norm": 0.3955257011722223, "learning_rate": 5.666666666666667e-06, "loss": 2.493, "step": 17 }, { "epoch": 0.058919803600654665, "grad_norm": 0.3954725948519232, "learning_rate": 6e-06, "loss": 2.4947, "step": 18 }, { "epoch": 0.062193126022913256, "grad_norm": 0.37175019309525287, "learning_rate": 6.333333333333333e-06, "loss": 2.5092, "step": 19 }, { "epoch": 0.06546644844517185, "grad_norm": 0.39057048150465146, "learning_rate": 6.666666666666667e-06, "loss": 2.4981, "step": 20 }, { "epoch": 0.06873977086743044, "grad_norm": 0.39262665115560863, "learning_rate": 7e-06, "loss": 2.4854, "step": 21 }, { "epoch": 0.07201309328968904, "grad_norm": 0.3951637407824461, "learning_rate": 7.333333333333333e-06, "loss": 2.5021, "step": 22 }, { "epoch": 0.07528641571194762, "grad_norm": 0.40140742815376235, "learning_rate": 7.666666666666667e-06, "loss": 2.517, "step": 23 }, { "epoch": 0.07855973813420622, "grad_norm": 0.36348786250077453, "learning_rate": 8.000000000000001e-06, "loss": 2.492, "step": 24 }, { "epoch": 0.08183306055646482, "grad_norm": 0.3927422984780222, "learning_rate": 8.333333333333334e-06, "loss": 2.4891, "step": 25 }, { "epoch": 0.0851063829787234, "grad_norm": 0.3802256129939313, "learning_rate": 8.666666666666668e-06, "loss": 2.5241, "step": 26 }, { "epoch": 0.088379705400982, "grad_norm": 0.4048560033284272, "learning_rate": 9e-06, "loss": 2.4996, "step": 27 }, { "epoch": 0.09165302782324058, "grad_norm": 0.3938035425880514, "learning_rate": 9.333333333333334e-06, "loss": 2.51, "step": 28 }, { "epoch": 0.09492635024549918, "grad_norm": 0.4318232114640702, "learning_rate": 9.666666666666667e-06, "loss": 2.5345, "step": 29 }, { "epoch": 0.09819967266775777, "grad_norm": 0.3960154579660599, "learning_rate": 1e-05, "loss": 2.511, "step": 30 }, { "epoch": 0.10147299509001637, "grad_norm": 0.4569501622130681, "learning_rate": 9.999673735634259e-06, "loss": 2.4827, "step": 31 }, { "epoch": 0.10474631751227496, "grad_norm": 0.3958075272764396, "learning_rate": 9.998694985116406e-06, "loss": 2.4884, "step": 32 }, { "epoch": 0.10801963993453355, "grad_norm": 0.37500089768123596, "learning_rate": 9.997063876179007e-06, "loss": 2.5027, "step": 33 }, { "epoch": 0.11129296235679215, "grad_norm": 0.3815772878145323, "learning_rate": 9.994780621691156e-06, "loss": 2.4957, "step": 34 }, { "epoch": 0.11456628477905073, "grad_norm": 0.38634081421059385, "learning_rate": 9.991845519630679e-06, "loss": 2.4937, "step": 35 }, { "epoch": 0.11783960720130933, "grad_norm": 0.3671354090721152, "learning_rate": 9.988258953045264e-06, "loss": 2.4789, "step": 36 }, { "epoch": 0.12111292962356793, "grad_norm": 0.36190374114193224, "learning_rate": 9.984021390002458e-06, "loss": 2.499, "step": 37 }, { "epoch": 0.12438625204582651, "grad_norm": 0.38466932485680505, "learning_rate": 9.979133383528591e-06, "loss": 2.5026, "step": 38 }, { "epoch": 0.1276595744680851, "grad_norm": 0.3573078258794223, "learning_rate": 9.973595571536593e-06, "loss": 2.4787, "step": 39 }, { "epoch": 0.1309328968903437, "grad_norm": 0.39354137647528986, "learning_rate": 9.96740867674275e-06, "loss": 2.5228, "step": 40 }, { "epoch": 0.1342062193126023, "grad_norm": 0.3834121962923347, "learning_rate": 9.960573506572391e-06, "loss": 2.5147, "step": 41 }, { "epoch": 0.13747954173486088, "grad_norm": 0.383218733958223, "learning_rate": 9.953090953054491e-06, "loss": 2.4847, "step": 42 }, { "epoch": 0.1407528641571195, "grad_norm": 0.3964524303138896, "learning_rate": 9.944961992705288e-06, "loss": 2.4953, "step": 43 }, { "epoch": 0.14402618657937807, "grad_norm": 0.3907082692537958, "learning_rate": 9.936187686400814e-06, "loss": 2.5074, "step": 44 }, { "epoch": 0.14729950900163666, "grad_norm": 0.42441529396911143, "learning_rate": 9.926769179238467e-06, "loss": 2.4694, "step": 45 }, { "epoch": 0.15057283142389524, "grad_norm": 0.36309886081053155, "learning_rate": 9.916707700387546e-06, "loss": 2.4816, "step": 46 }, { "epoch": 0.15384615384615385, "grad_norm": 0.3745053331706986, "learning_rate": 9.906004562928865e-06, "loss": 2.4876, "step": 47 }, { "epoch": 0.15711947626841244, "grad_norm": 0.3558032864873746, "learning_rate": 9.894661163683361e-06, "loss": 2.4887, "step": 48 }, { "epoch": 0.16039279869067102, "grad_norm": 0.3560175097976504, "learning_rate": 9.882678983029819e-06, "loss": 2.4817, "step": 49 }, { "epoch": 0.16366612111292964, "grad_norm": 0.36618662107001704, "learning_rate": 9.870059584711668e-06, "loss": 2.4665, "step": 50 }, { "epoch": 0.16693944353518822, "grad_norm": 0.3601957684934551, "learning_rate": 9.856804615632904e-06, "loss": 2.4854, "step": 51 }, { "epoch": 0.1702127659574468, "grad_norm": 0.3704954734907262, "learning_rate": 9.842915805643156e-06, "loss": 2.5347, "step": 52 }, { "epoch": 0.1734860883797054, "grad_norm": 0.36959475458440244, "learning_rate": 9.82839496731194e-06, "loss": 2.4887, "step": 53 }, { "epoch": 0.176759410801964, "grad_norm": 0.384552486477283, "learning_rate": 9.813243995692097e-06, "loss": 2.4753, "step": 54 }, { "epoch": 0.18003273322422259, "grad_norm": 0.36249860452124605, "learning_rate": 9.797464868072489e-06, "loss": 2.469, "step": 55 }, { "epoch": 0.18330605564648117, "grad_norm": 0.3648846542235022, "learning_rate": 9.781059643719937e-06, "loss": 2.4984, "step": 56 }, { "epoch": 0.18657937806873978, "grad_norm": 0.37334256629644896, "learning_rate": 9.76403046361049e-06, "loss": 2.456, "step": 57 }, { "epoch": 0.18985270049099837, "grad_norm": 0.36980026599186505, "learning_rate": 9.74637955015001e-06, "loss": 2.485, "step": 58 }, { "epoch": 0.19312602291325695, "grad_norm": 0.40834498386413537, "learning_rate": 9.728109206884125e-06, "loss": 2.5011, "step": 59 }, { "epoch": 0.19639934533551553, "grad_norm": 0.35772757844310776, "learning_rate": 9.709221818197626e-06, "loss": 2.4915, "step": 60 }, { "epoch": 0.19967266775777415, "grad_norm": 0.5331894263162329, "learning_rate": 9.689719849003261e-06, "loss": 2.4779, "step": 61 }, { "epoch": 0.20294599018003273, "grad_norm": 0.5108742054472333, "learning_rate": 9.66960584442008e-06, "loss": 2.5002, "step": 62 }, { "epoch": 0.20621931260229132, "grad_norm": 0.36603499456334015, "learning_rate": 9.648882429441258e-06, "loss": 2.4605, "step": 63 }, { "epoch": 0.20949263502454993, "grad_norm": 0.3858608557671342, "learning_rate": 9.627552308591534e-06, "loss": 2.484, "step": 64 }, { "epoch": 0.2127659574468085, "grad_norm": 0.39024218298879865, "learning_rate": 9.60561826557425e-06, "loss": 2.5057, "step": 65 }, { "epoch": 0.2160392798690671, "grad_norm": 0.38798042222376583, "learning_rate": 9.58308316290806e-06, "loss": 2.4573, "step": 66 }, { "epoch": 0.2193126022913257, "grad_norm": 0.4495343758022552, "learning_rate": 9.559949941553351e-06, "loss": 2.4751, "step": 67 }, { "epoch": 0.2225859247135843, "grad_norm": 0.5550558249858909, "learning_rate": 9.536221620528442e-06, "loss": 2.4706, "step": 68 }, { "epoch": 0.22585924713584288, "grad_norm": 0.7753792348358921, "learning_rate": 9.511901296515578e-06, "loss": 2.4536, "step": 69 }, { "epoch": 0.22913256955810146, "grad_norm": 0.4620597211236889, "learning_rate": 9.486992143456792e-06, "loss": 2.4966, "step": 70 }, { "epoch": 0.23240589198036007, "grad_norm": 2.0294893276083674, "learning_rate": 9.461497412139697e-06, "loss": 2.4896, "step": 71 }, { "epoch": 0.23567921440261866, "grad_norm": 3.9731994256120258, "learning_rate": 9.435420429773227e-06, "loss": 2.4775, "step": 72 }, { "epoch": 0.23895253682487724, "grad_norm": 3.716661029391411, "learning_rate": 9.408764599553429e-06, "loss": 2.4774, "step": 73 }, { "epoch": 0.24222585924713586, "grad_norm": 2.4511968994240396, "learning_rate": 9.381533400219319e-06, "loss": 2.4545, "step": 74 }, { "epoch": 0.24549918166939444, "grad_norm": 0.9224443041076055, "learning_rate": 9.353730385598887e-06, "loss": 2.4927, "step": 75 }, { "epoch": 0.24877250409165302, "grad_norm": 0.4657968594732789, "learning_rate": 9.325359184145307e-06, "loss": 2.4565, "step": 76 }, { "epoch": 0.2520458265139116, "grad_norm": 0.7177608849321568, "learning_rate": 9.296423498463396e-06, "loss": 2.4825, "step": 77 }, { "epoch": 0.2553191489361702, "grad_norm": 0.57974277665311, "learning_rate": 9.26692710482641e-06, "loss": 2.4789, "step": 78 }, { "epoch": 0.25859247135842883, "grad_norm": 0.5116606332293397, "learning_rate": 9.236873852683213e-06, "loss": 2.4654, "step": 79 }, { "epoch": 0.2618657937806874, "grad_norm": 0.4372804881120678, "learning_rate": 9.206267664155906e-06, "loss": 2.4649, "step": 80 }, { "epoch": 0.265139116202946, "grad_norm": 0.4185499524749441, "learning_rate": 9.175112533527963e-06, "loss": 2.5026, "step": 81 }, { "epoch": 0.2684124386252046, "grad_norm": 0.5272610155007883, "learning_rate": 9.143412526722958e-06, "loss": 2.4975, "step": 82 }, { "epoch": 0.27168576104746317, "grad_norm": 0.5340935332042765, "learning_rate": 9.111171780773938e-06, "loss": 2.4816, "step": 83 }, { "epoch": 0.27495908346972175, "grad_norm": 0.3904804241846333, "learning_rate": 9.078394503283509e-06, "loss": 2.4675, "step": 84 }, { "epoch": 0.27823240589198034, "grad_norm": 0.3328349798561506, "learning_rate": 9.045084971874738e-06, "loss": 2.4633, "step": 85 }, { "epoch": 0.281505728314239, "grad_norm": 0.37021724634968695, "learning_rate": 9.011247533632876e-06, "loss": 2.4508, "step": 86 }, { "epoch": 0.28477905073649756, "grad_norm": 0.3729529690097639, "learning_rate": 8.976886604538055e-06, "loss": 2.489, "step": 87 }, { "epoch": 0.28805237315875615, "grad_norm": 0.33761554067347327, "learning_rate": 8.942006668888972e-06, "loss": 2.4734, "step": 88 }, { "epoch": 0.29132569558101473, "grad_norm": 0.3584893895881646, "learning_rate": 8.906612278717657e-06, "loss": 2.4702, "step": 89 }, { "epoch": 0.2945990180032733, "grad_norm": 0.3546249378424476, "learning_rate": 8.870708053195414e-06, "loss": 2.4891, "step": 90 }, { "epoch": 0.2978723404255319, "grad_norm": 0.3472403078226491, "learning_rate": 8.834298678029988e-06, "loss": 2.4671, "step": 91 }, { "epoch": 0.3011456628477905, "grad_norm": 0.3453678794183207, "learning_rate": 8.797388904854064e-06, "loss": 2.4971, "step": 92 }, { "epoch": 0.3044189852700491, "grad_norm": 0.35760827827717245, "learning_rate": 8.759983550605132e-06, "loss": 2.478, "step": 93 }, { "epoch": 0.3076923076923077, "grad_norm": 0.32637961131906473, "learning_rate": 8.72208749689686e-06, "loss": 2.4519, "step": 94 }, { "epoch": 0.3109656301145663, "grad_norm": 0.3451716663724565, "learning_rate": 8.683705689382025e-06, "loss": 2.4597, "step": 95 }, { "epoch": 0.3142389525368249, "grad_norm": 0.3397179170708975, "learning_rate": 8.644843137107058e-06, "loss": 2.4738, "step": 96 }, { "epoch": 0.31751227495908346, "grad_norm": 0.3507051892225581, "learning_rate": 8.605504911858347e-06, "loss": 2.5, "step": 97 }, { "epoch": 0.32078559738134205, "grad_norm": 0.3400362404421244, "learning_rate": 8.565696147500338e-06, "loss": 2.4768, "step": 98 }, { "epoch": 0.32405891980360063, "grad_norm": 0.35484798538066187, "learning_rate": 8.525422039305529e-06, "loss": 2.4858, "step": 99 }, { "epoch": 0.32733224222585927, "grad_norm": 0.3658788522731373, "learning_rate": 8.48468784327647e-06, "loss": 2.4738, "step": 100 }, { "epoch": 0.33060556464811786, "grad_norm": 0.3398517081347872, "learning_rate": 8.44349887545981e-06, "loss": 2.4877, "step": 101 }, { "epoch": 0.33387888707037644, "grad_norm": 0.38013549611094194, "learning_rate": 8.401860511252535e-06, "loss": 2.4638, "step": 102 }, { "epoch": 0.337152209492635, "grad_norm": 0.3484658452958359, "learning_rate": 8.35977818470044e-06, "loss": 2.4783, "step": 103 }, { "epoch": 0.3404255319148936, "grad_norm": 0.3554237638577351, "learning_rate": 8.31725738778896e-06, "loss": 2.474, "step": 104 }, { "epoch": 0.3436988543371522, "grad_norm": 0.3448509168732586, "learning_rate": 8.274303669726427e-06, "loss": 2.4568, "step": 105 }, { "epoch": 0.3469721767594108, "grad_norm": 0.3491512405342342, "learning_rate": 8.230922636219872e-06, "loss": 2.479, "step": 106 }, { "epoch": 0.3502454991816694, "grad_norm": 0.35293167144622156, "learning_rate": 8.18711994874345e-06, "loss": 2.4889, "step": 107 }, { "epoch": 0.353518821603928, "grad_norm": 0.34554222359891335, "learning_rate": 8.142901323799578e-06, "loss": 2.4948, "step": 108 }, { "epoch": 0.3567921440261866, "grad_norm": 0.3433706276706439, "learning_rate": 8.098272532172906e-06, "loss": 2.4896, "step": 109 }, { "epoch": 0.36006546644844517, "grad_norm": 0.32061609152441495, "learning_rate": 8.053239398177191e-06, "loss": 2.4374, "step": 110 }, { "epoch": 0.36333878887070375, "grad_norm": 0.3443744364995488, "learning_rate": 8.007807798895195e-06, "loss": 2.4754, "step": 111 }, { "epoch": 0.36661211129296234, "grad_norm": 0.32691985158573217, "learning_rate": 7.961983663411684e-06, "loss": 2.4777, "step": 112 }, { "epoch": 0.3698854337152209, "grad_norm": 0.33245765943085837, "learning_rate": 7.91577297203966e-06, "loss": 2.4723, "step": 113 }, { "epoch": 0.37315875613747956, "grad_norm": 0.33707956489119834, "learning_rate": 7.869181755539888e-06, "loss": 2.4709, "step": 114 }, { "epoch": 0.37643207855973815, "grad_norm": 0.3334493326179338, "learning_rate": 7.822216094333847e-06, "loss": 2.5068, "step": 115 }, { "epoch": 0.37970540098199673, "grad_norm": 0.3244946372709713, "learning_rate": 7.774882117710203e-06, "loss": 2.4797, "step": 116 }, { "epoch": 0.3829787234042553, "grad_norm": 0.3277650850109358, "learning_rate": 7.727186003024902e-06, "loss": 2.4618, "step": 117 }, { "epoch": 0.3862520458265139, "grad_norm": 0.34009316740947815, "learning_rate": 7.679133974894984e-06, "loss": 2.4581, "step": 118 }, { "epoch": 0.3895253682487725, "grad_norm": 0.3381461847729053, "learning_rate": 7.630732304386244e-06, "loss": 2.4817, "step": 119 }, { "epoch": 0.39279869067103107, "grad_norm": 0.3385097066946103, "learning_rate": 7.5819873081948105e-06, "loss": 2.4847, "step": 120 }, { "epoch": 0.3960720130932897, "grad_norm": 0.3372871065877283, "learning_rate": 7.532905347822792e-06, "loss": 2.4577, "step": 121 }, { "epoch": 0.3993453355155483, "grad_norm": 0.32559796645917616, "learning_rate": 7.4834928287480566e-06, "loss": 2.4923, "step": 122 }, { "epoch": 0.4026186579378069, "grad_norm": 0.3251819544833203, "learning_rate": 7.433756199588282e-06, "loss": 2.4746, "step": 123 }, { "epoch": 0.40589198036006546, "grad_norm": 0.33165207760018856, "learning_rate": 7.383701951259375e-06, "loss": 2.497, "step": 124 }, { "epoch": 0.40916530278232405, "grad_norm": 0.33880444150123695, "learning_rate": 7.333336616128369e-06, "loss": 2.4362, "step": 125 }, { "epoch": 0.41243862520458263, "grad_norm": 0.32428200516815325, "learning_rate": 7.282666767160913e-06, "loss": 2.4439, "step": 126 }, { "epoch": 0.4157119476268412, "grad_norm": 0.3273982952564575, "learning_rate": 7.23169901706346e-06, "loss": 2.4801, "step": 127 }, { "epoch": 0.41898527004909986, "grad_norm": 0.3261565580453942, "learning_rate": 7.180440017420277e-06, "loss": 2.4903, "step": 128 }, { "epoch": 0.42225859247135844, "grad_norm": 0.35528485781883334, "learning_rate": 7.128896457825364e-06, "loss": 2.4653, "step": 129 }, { "epoch": 0.425531914893617, "grad_norm": 0.3185186205453131, "learning_rate": 7.0770750650094335e-06, "loss": 2.4605, "step": 130 }, { "epoch": 0.4288052373158756, "grad_norm": 0.3431153323420711, "learning_rate": 7.024982601962027e-06, "loss": 2.4627, "step": 131 }, { "epoch": 0.4320785597381342, "grad_norm": 0.31409563060102685, "learning_rate": 6.972625867048914e-06, "loss": 2.457, "step": 132 }, { "epoch": 0.4353518821603928, "grad_norm": 0.3352629031883897, "learning_rate": 6.9200116931248575e-06, "loss": 2.4805, "step": 133 }, { "epoch": 0.4386252045826514, "grad_norm": 0.3295764773687296, "learning_rate": 6.8671469466418914e-06, "loss": 2.465, "step": 134 }, { "epoch": 0.44189852700491, "grad_norm": 0.32638616945204396, "learning_rate": 6.814038526753205e-06, "loss": 2.4317, "step": 135 }, { "epoch": 0.4451718494271686, "grad_norm": 0.358055738131873, "learning_rate": 6.760693364412776e-06, "loss": 2.4459, "step": 136 }, { "epoch": 0.44844517184942717, "grad_norm": 0.32782232063682554, "learning_rate": 6.707118421470822e-06, "loss": 2.4652, "step": 137 }, { "epoch": 0.45171849427168576, "grad_norm": 0.3213728199970128, "learning_rate": 6.653320689765257e-06, "loss": 2.4812, "step": 138 }, { "epoch": 0.45499181669394434, "grad_norm": 0.3145652401211947, "learning_rate": 6.599307190209206e-06, "loss": 2.4622, "step": 139 }, { "epoch": 0.4582651391162029, "grad_norm": 0.33107965983633225, "learning_rate": 6.545084971874738e-06, "loss": 2.4856, "step": 140 }, { "epoch": 0.46153846153846156, "grad_norm": 0.33391686920613367, "learning_rate": 6.490661111072923e-06, "loss": 2.4519, "step": 141 }, { "epoch": 0.46481178396072015, "grad_norm": 0.31915606377667516, "learning_rate": 6.4360427104303326e-06, "loss": 2.458, "step": 142 }, { "epoch": 0.46808510638297873, "grad_norm": 0.32619864635648854, "learning_rate": 6.381236897962102e-06, "loss": 2.48, "step": 143 }, { "epoch": 0.4713584288052373, "grad_norm": 0.3344755693427373, "learning_rate": 6.326250826141689e-06, "loss": 2.4728, "step": 144 }, { "epoch": 0.4746317512274959, "grad_norm": 0.3125523770454477, "learning_rate": 6.271091670967437e-06, "loss": 2.4587, "step": 145 }, { "epoch": 0.4779050736497545, "grad_norm": 0.32456956099182205, "learning_rate": 6.215766631026049e-06, "loss": 2.4694, "step": 146 }, { "epoch": 0.48117839607201307, "grad_norm": 0.3183205037717091, "learning_rate": 6.1602829265531585e-06, "loss": 2.4752, "step": 147 }, { "epoch": 0.4844517184942717, "grad_norm": 0.31706883694319504, "learning_rate": 6.1046477984910215e-06, "loss": 2.4715, "step": 148 }, { "epoch": 0.4877250409165303, "grad_norm": 0.32677657711179986, "learning_rate": 6.048868507543547e-06, "loss": 2.4684, "step": 149 }, { "epoch": 0.4909983633387889, "grad_norm": 0.32457051719454905, "learning_rate": 5.9929523332287275e-06, "loss": 2.471, "step": 150 }, { "epoch": 0.49427168576104746, "grad_norm": 0.329167055323462, "learning_rate": 5.936906572928625e-06, "loss": 2.4696, "step": 151 }, { "epoch": 0.49754500818330605, "grad_norm": 0.3233867081897887, "learning_rate": 5.880738540937008e-06, "loss": 2.4758, "step": 152 }, { "epoch": 0.5008183306055647, "grad_norm": 0.310943717613368, "learning_rate": 5.824455567504817e-06, "loss": 2.4813, "step": 153 }, { "epoch": 0.5040916530278232, "grad_norm": 0.35767604251132745, "learning_rate": 5.7680649978834976e-06, "loss": 2.4628, "step": 154 }, { "epoch": 0.5073649754500819, "grad_norm": 0.3066334220540355, "learning_rate": 5.711574191366427e-06, "loss": 2.4645, "step": 155 }, { "epoch": 0.5106382978723404, "grad_norm": 0.3299515321649155, "learning_rate": 5.654990520328465e-06, "loss": 2.4445, "step": 156 }, { "epoch": 0.513911620294599, "grad_norm": 0.31272357276253326, "learning_rate": 5.59832136926383e-06, "loss": 2.4594, "step": 157 }, { "epoch": 0.5171849427168577, "grad_norm": 0.3094045826418548, "learning_rate": 5.541574133822374e-06, "loss": 2.4481, "step": 158 }, { "epoch": 0.5204582651391162, "grad_norm": 0.33143709507448227, "learning_rate": 5.484756219844408e-06, "loss": 2.4621, "step": 159 }, { "epoch": 0.5237315875613748, "grad_norm": 0.30398134753496064, "learning_rate": 5.4278750423942e-06, "loss": 2.4715, "step": 160 }, { "epoch": 0.5270049099836334, "grad_norm": 0.31511994044804714, "learning_rate": 5.370938024792262e-06, "loss": 2.4609, "step": 161 }, { "epoch": 0.530278232405892, "grad_norm": 0.3204336918739602, "learning_rate": 5.3139525976465675e-06, "loss": 2.4501, "step": 162 }, { "epoch": 0.5335515548281505, "grad_norm": 0.3243453056297712, "learning_rate": 5.2569261978828155e-06, "loss": 2.4634, "step": 163 }, { "epoch": 0.5368248772504092, "grad_norm": 0.3131904676215095, "learning_rate": 5.199866267773868e-06, "loss": 2.4684, "step": 164 }, { "epoch": 0.5400981996726678, "grad_norm": 0.3338712281774405, "learning_rate": 5.142780253968481e-06, "loss": 2.4687, "step": 165 }, { "epoch": 0.5433715220949263, "grad_norm": 0.3108967805240922, "learning_rate": 5.085675606519496e-06, "loss": 2.4443, "step": 166 }, { "epoch": 0.546644844517185, "grad_norm": 0.31028829982787426, "learning_rate": 5.028559777911543e-06, "loss": 2.471, "step": 167 }, { "epoch": 0.5499181669394435, "grad_norm": 0.31679487774858733, "learning_rate": 4.971440222088459e-06, "loss": 2.4654, "step": 168 }, { "epoch": 0.5531914893617021, "grad_norm": 0.31296921041810605, "learning_rate": 4.914324393480504e-06, "loss": 2.4442, "step": 169 }, { "epoch": 0.5564648117839607, "grad_norm": 0.3165078522607312, "learning_rate": 4.85721974603152e-06, "loss": 2.4494, "step": 170 }, { "epoch": 0.5597381342062193, "grad_norm": 0.3540062454779396, "learning_rate": 4.800133732226135e-06, "loss": 2.4664, "step": 171 }, { "epoch": 0.563011456628478, "grad_norm": 0.32009750544034943, "learning_rate": 4.743073802117185e-06, "loss": 2.4809, "step": 172 }, { "epoch": 0.5662847790507365, "grad_norm": 0.2983415714217039, "learning_rate": 4.686047402353433e-06, "loss": 2.4564, "step": 173 }, { "epoch": 0.5695581014729951, "grad_norm": 0.2999034797032862, "learning_rate": 4.62906197520774e-06, "loss": 2.467, "step": 174 }, { "epoch": 0.5728314238952537, "grad_norm": 0.3102585042911049, "learning_rate": 4.572124957605803e-06, "loss": 2.46, "step": 175 }, { "epoch": 0.5761047463175123, "grad_norm": 0.31057870076393, "learning_rate": 4.515243780155594e-06, "loss": 2.4704, "step": 176 }, { "epoch": 0.5793780687397708, "grad_norm": 0.3046118860204264, "learning_rate": 4.458425866177628e-06, "loss": 2.467, "step": 177 }, { "epoch": 0.5826513911620295, "grad_norm": 0.30073951913953495, "learning_rate": 4.401678630736172e-06, "loss": 2.4743, "step": 178 }, { "epoch": 0.5859247135842881, "grad_norm": 0.30406830972525584, "learning_rate": 4.3450094796715354e-06, "loss": 2.4798, "step": 179 }, { "epoch": 0.5891980360065466, "grad_norm": 0.3054725070167328, "learning_rate": 4.2884258086335755e-06, "loss": 2.468, "step": 180 }, { "epoch": 0.5924713584288053, "grad_norm": 0.31961569501005616, "learning_rate": 4.231935002116504e-06, "loss": 2.4853, "step": 181 }, { "epoch": 0.5957446808510638, "grad_norm": 0.29332706934686226, "learning_rate": 4.175544432495184e-06, "loss": 2.4684, "step": 182 }, { "epoch": 0.5990180032733224, "grad_norm": 0.3136548340197598, "learning_rate": 4.119261459062992e-06, "loss": 2.4617, "step": 183 }, { "epoch": 0.602291325695581, "grad_norm": 0.29768754171106776, "learning_rate": 4.063093427071376e-06, "loss": 2.4506, "step": 184 }, { "epoch": 0.6055646481178396, "grad_norm": 0.3102339131602259, "learning_rate": 4.007047666771274e-06, "loss": 2.4692, "step": 185 }, { "epoch": 0.6088379705400983, "grad_norm": 0.31512899490271845, "learning_rate": 3.951131492456455e-06, "loss": 2.4399, "step": 186 }, { "epoch": 0.6121112929623568, "grad_norm": 0.29668620197615053, "learning_rate": 3.895352201508981e-06, "loss": 2.4658, "step": 187 }, { "epoch": 0.6153846153846154, "grad_norm": 0.3129729278174446, "learning_rate": 3.839717073446842e-06, "loss": 2.4296, "step": 188 }, { "epoch": 0.618657937806874, "grad_norm": 0.3087654037315499, "learning_rate": 3.7842333689739524e-06, "loss": 2.4512, "step": 189 }, { "epoch": 0.6219312602291326, "grad_norm": 0.29989373586590584, "learning_rate": 3.7289083290325668e-06, "loss": 2.4543, "step": 190 }, { "epoch": 0.6252045826513911, "grad_norm": 0.30200062774615277, "learning_rate": 3.673749173858312e-06, "loss": 2.4637, "step": 191 }, { "epoch": 0.6284779050736498, "grad_norm": 0.3059048859940154, "learning_rate": 3.618763102037899e-06, "loss": 2.4714, "step": 192 }, { "epoch": 0.6317512274959084, "grad_norm": 0.2995831991905797, "learning_rate": 3.563957289569669e-06, "loss": 2.4581, "step": 193 }, { "epoch": 0.6350245499181669, "grad_norm": 0.3058506564330453, "learning_rate": 3.509338888927079e-06, "loss": 2.4556, "step": 194 }, { "epoch": 0.6382978723404256, "grad_norm": 0.2932520015994215, "learning_rate": 3.4549150281252635e-06, "loss": 2.4536, "step": 195 }, { "epoch": 0.6415711947626841, "grad_norm": 0.29644091797619826, "learning_rate": 3.400692809790796e-06, "loss": 2.4524, "step": 196 }, { "epoch": 0.6448445171849427, "grad_norm": 0.2973231753014199, "learning_rate": 3.346679310234744e-06, "loss": 2.439, "step": 197 }, { "epoch": 0.6481178396072013, "grad_norm": 0.2970530949566317, "learning_rate": 3.292881578529179e-06, "loss": 2.441, "step": 198 }, { "epoch": 0.6513911620294599, "grad_norm": 0.29457945074596364, "learning_rate": 3.2393066355872264e-06, "loss": 2.4393, "step": 199 }, { "epoch": 0.6546644844517185, "grad_norm": 0.2959752353295032, "learning_rate": 3.1859614732467957e-06, "loss": 2.4413, "step": 200 }, { "epoch": 0.6579378068739771, "grad_norm": 0.2894024794008193, "learning_rate": 3.1328530533581102e-06, "loss": 2.4486, "step": 201 }, { "epoch": 0.6612111292962357, "grad_norm": 0.30092182634576625, "learning_rate": 3.0799883068751433e-06, "loss": 2.4523, "step": 202 }, { "epoch": 0.6644844517184942, "grad_norm": 0.2957309468597844, "learning_rate": 3.0273741329510852e-06, "loss": 2.4335, "step": 203 }, { "epoch": 0.6677577741407529, "grad_norm": 0.28459255910232356, "learning_rate": 2.975017398037974e-06, "loss": 2.4875, "step": 204 }, { "epoch": 0.6710310965630114, "grad_norm": 0.29271515123042124, "learning_rate": 2.9229249349905686e-06, "loss": 2.475, "step": 205 }, { "epoch": 0.67430441898527, "grad_norm": 0.30277820836046737, "learning_rate": 2.871103542174637e-06, "loss": 2.4669, "step": 206 }, { "epoch": 0.6775777414075287, "grad_norm": 0.2859827817423741, "learning_rate": 2.8195599825797233e-06, "loss": 2.4514, "step": 207 }, { "epoch": 0.6808510638297872, "grad_norm": 0.2968379809937716, "learning_rate": 2.7683009829365417e-06, "loss": 2.4607, "step": 208 }, { "epoch": 0.6841243862520459, "grad_norm": 0.29656383723002844, "learning_rate": 2.717333232839088e-06, "loss": 2.4905, "step": 209 }, { "epoch": 0.6873977086743044, "grad_norm": 0.28883812040540374, "learning_rate": 2.6666633838716317e-06, "loss": 2.4543, "step": 210 }, { "epoch": 0.690671031096563, "grad_norm": 0.29531117152054726, "learning_rate": 2.616298048740626e-06, "loss": 2.4551, "step": 211 }, { "epoch": 0.6939443535188216, "grad_norm": 0.2862187454441366, "learning_rate": 2.566243800411719e-06, "loss": 2.452, "step": 212 }, { "epoch": 0.6972176759410802, "grad_norm": 0.2911286270146174, "learning_rate": 2.5165071712519447e-06, "loss": 2.4624, "step": 213 }, { "epoch": 0.7004909983633388, "grad_norm": 0.2869776645891394, "learning_rate": 2.467094652177209e-06, "loss": 2.4607, "step": 214 }, { "epoch": 0.7037643207855974, "grad_norm": 0.2811233703409396, "learning_rate": 2.418012691805191e-06, "loss": 2.4739, "step": 215 }, { "epoch": 0.707037643207856, "grad_norm": 0.28005066226487846, "learning_rate": 2.3692676956137585e-06, "loss": 2.457, "step": 216 }, { "epoch": 0.7103109656301145, "grad_norm": 0.29229180839794044, "learning_rate": 2.320866025105016e-06, "loss": 2.4505, "step": 217 }, { "epoch": 0.7135842880523732, "grad_norm": 0.28905509654195427, "learning_rate": 2.2728139969751005e-06, "loss": 2.4497, "step": 218 }, { "epoch": 0.7168576104746317, "grad_norm": 0.29091875602679856, "learning_rate": 2.225117882289799e-06, "loss": 2.4386, "step": 219 }, { "epoch": 0.7201309328968903, "grad_norm": 0.28132598389097907, "learning_rate": 2.1777839056661555e-06, "loss": 2.4688, "step": 220 }, { "epoch": 0.723404255319149, "grad_norm": 0.27552653164791197, "learning_rate": 2.1308182444601126e-06, "loss": 2.4515, "step": 221 }, { "epoch": 0.7266775777414075, "grad_norm": 0.285817433950407, "learning_rate": 2.0842270279603403e-06, "loss": 2.4321, "step": 222 }, { "epoch": 0.7299509001636661, "grad_norm": 0.2774842353710076, "learning_rate": 2.0380163365883188e-06, "loss": 2.4686, "step": 223 }, { "epoch": 0.7332242225859247, "grad_norm": 0.2782616476099915, "learning_rate": 1.9921922011048065e-06, "loss": 2.4438, "step": 224 }, { "epoch": 0.7364975450081833, "grad_norm": 0.27697006620218906, "learning_rate": 1.946760601822809e-06, "loss": 2.4461, "step": 225 }, { "epoch": 0.7397708674304418, "grad_norm": 0.2794520367771161, "learning_rate": 1.9017274678270948e-06, "loss": 2.457, "step": 226 }, { "epoch": 0.7430441898527005, "grad_norm": 0.293225655343382, "learning_rate": 1.8570986762004246e-06, "loss": 2.4339, "step": 227 }, { "epoch": 0.7463175122749591, "grad_norm": 0.28410174489532863, "learning_rate": 1.8128800512565514e-06, "loss": 2.4398, "step": 228 }, { "epoch": 0.7495908346972177, "grad_norm": 0.28113616787582135, "learning_rate": 1.7690773637801295e-06, "loss": 2.463, "step": 229 }, { "epoch": 0.7528641571194763, "grad_norm": 0.27128743447878584, "learning_rate": 1.7256963302735752e-06, "loss": 2.4638, "step": 230 }, { "epoch": 0.7561374795417348, "grad_norm": 0.28051379395375337, "learning_rate": 1.6827426122110412e-06, "loss": 2.4803, "step": 231 }, { "epoch": 0.7594108019639935, "grad_norm": 0.2774673111470984, "learning_rate": 1.6402218152995609e-06, "loss": 2.4255, "step": 232 }, { "epoch": 0.762684124386252, "grad_norm": 0.26906487455817846, "learning_rate": 1.598139488747467e-06, "loss": 2.4478, "step": 233 }, { "epoch": 0.7659574468085106, "grad_norm": 0.2764519019580199, "learning_rate": 1.5565011245401928e-06, "loss": 2.4674, "step": 234 }, { "epoch": 0.7692307692307693, "grad_norm": 0.28254452300968197, "learning_rate": 1.5153121567235334e-06, "loss": 2.4648, "step": 235 }, { "epoch": 0.7725040916530278, "grad_norm": 0.28044552819552365, "learning_rate": 1.4745779606944716e-06, "loss": 2.4634, "step": 236 }, { "epoch": 0.7757774140752864, "grad_norm": 0.2883807662161686, "learning_rate": 1.4343038524996645e-06, "loss": 2.4409, "step": 237 }, { "epoch": 0.779050736497545, "grad_norm": 0.2712743023254104, "learning_rate": 1.3944950881416541e-06, "loss": 2.4749, "step": 238 }, { "epoch": 0.7823240589198036, "grad_norm": 0.2844283835642021, "learning_rate": 1.3551568628929434e-06, "loss": 2.4682, "step": 239 }, { "epoch": 0.7855973813420621, "grad_norm": 0.2792759057607264, "learning_rate": 1.3162943106179748e-06, "loss": 2.4414, "step": 240 }, { "epoch": 0.7888707037643208, "grad_norm": 0.2733989225438821, "learning_rate": 1.2779125031031413e-06, "loss": 2.4257, "step": 241 }, { "epoch": 0.7921440261865794, "grad_norm": 0.271099884675861, "learning_rate": 1.2400164493948713e-06, "loss": 2.4569, "step": 242 }, { "epoch": 0.795417348608838, "grad_norm": 0.2745676234993269, "learning_rate": 1.2026110951459364e-06, "loss": 2.4489, "step": 243 }, { "epoch": 0.7986906710310966, "grad_norm": 0.2767138934905047, "learning_rate": 1.1657013219700108e-06, "loss": 2.4562, "step": 244 }, { "epoch": 0.8019639934533551, "grad_norm": 0.2742378409963404, "learning_rate": 1.1292919468045876e-06, "loss": 2.4333, "step": 245 }, { "epoch": 0.8052373158756138, "grad_norm": 0.2585166236299042, "learning_rate": 1.0933877212823462e-06, "loss": 2.4476, "step": 246 }, { "epoch": 0.8085106382978723, "grad_norm": 0.2696183429586083, "learning_rate": 1.057993331111029e-06, "loss": 2.4402, "step": 247 }, { "epoch": 0.8117839607201309, "grad_norm": 0.2742359647860916, "learning_rate": 1.0231133954619449e-06, "loss": 2.431, "step": 248 }, { "epoch": 0.8150572831423896, "grad_norm": 0.2745250582957182, "learning_rate": 9.887524663671243e-07, "loss": 2.456, "step": 249 }, { "epoch": 0.8183306055646481, "grad_norm": 0.2915439246328324, "learning_rate": 9.549150281252633e-07, "loss": 2.4739, "step": 250 }, { "epoch": 0.8216039279869067, "grad_norm": 0.2703132458554063, "learning_rate": 9.216054967164916e-07, "loss": 2.461, "step": 251 }, { "epoch": 0.8248772504091653, "grad_norm": 0.2902377178678097, "learning_rate": 8.888282192260645e-07, "loss": 2.4499, "step": 252 }, { "epoch": 0.8281505728314239, "grad_norm": 0.2732853452830904, "learning_rate": 8.565874732770429e-07, "loss": 2.4597, "step": 253 }, { "epoch": 0.8314238952536824, "grad_norm": 0.2696194963576079, "learning_rate": 8.248874664720375e-07, "loss": 2.4506, "step": 254 }, { "epoch": 0.8346972176759411, "grad_norm": 0.26860289744719607, "learning_rate": 7.937323358440935e-07, "loss": 2.4469, "step": 255 }, { "epoch": 0.8379705400981997, "grad_norm": 0.271233428425454, "learning_rate": 7.631261473167878e-07, "loss": 2.4532, "step": 256 }, { "epoch": 0.8412438625204582, "grad_norm": 0.2725978617713584, "learning_rate": 7.330728951735916e-07, "loss": 2.4745, "step": 257 }, { "epoch": 0.8445171849427169, "grad_norm": 0.26803651743311957, "learning_rate": 7.035765015366047e-07, "loss": 2.4381, "step": 258 }, { "epoch": 0.8477905073649754, "grad_norm": 0.25921598512129085, "learning_rate": 6.746408158546947e-07, "loss": 2.4533, "step": 259 }, { "epoch": 0.851063829787234, "grad_norm": 0.2733847496928928, "learning_rate": 6.462696144011149e-07, "loss": 2.4568, "step": 260 }, { "epoch": 0.8543371522094927, "grad_norm": 0.26178373020648743, "learning_rate": 6.184665997806832e-07, "loss": 2.4581, "step": 261 }, { "epoch": 0.8576104746317512, "grad_norm": 0.26098357909178693, "learning_rate": 5.912354004465709e-07, "loss": 2.4706, "step": 262 }, { "epoch": 0.8608837970540099, "grad_norm": 0.25697853000027987, "learning_rate": 5.645795702267731e-07, "loss": 2.4477, "step": 263 }, { "epoch": 0.8641571194762684, "grad_norm": 0.26192967323861704, "learning_rate": 5.385025878603039e-07, "loss": 2.4414, "step": 264 }, { "epoch": 0.867430441898527, "grad_norm": 0.26218888376989036, "learning_rate": 5.130078565432089e-07, "loss": 2.4518, "step": 265 }, { "epoch": 0.8707037643207856, "grad_norm": 0.26990782878286096, "learning_rate": 4.880987034844231e-07, "loss": 2.4553, "step": 266 }, { "epoch": 0.8739770867430442, "grad_norm": 0.26556563767940716, "learning_rate": 4.637783794715589e-07, "loss": 2.4513, "step": 267 }, { "epoch": 0.8772504091653028, "grad_norm": 0.2653970429292643, "learning_rate": 4.400500584466505e-07, "loss": 2.4545, "step": 268 }, { "epoch": 0.8805237315875614, "grad_norm": 0.26432968619903263, "learning_rate": 4.1691683709194184e-07, "loss": 2.4707, "step": 269 }, { "epoch": 0.88379705400982, "grad_norm": 0.2657171703652861, "learning_rate": 3.9438173442575e-07, "loss": 2.4651, "step": 270 }, { "epoch": 0.8870703764320785, "grad_norm": 0.2538156470759212, "learning_rate": 3.724476914084657e-07, "loss": 2.4449, "step": 271 }, { "epoch": 0.8903436988543372, "grad_norm": 0.27133692461664, "learning_rate": 3.511175705587433e-07, "loss": 2.473, "step": 272 }, { "epoch": 0.8936170212765957, "grad_norm": 0.25858521334517437, "learning_rate": 3.303941555799223e-07, "loss": 2.4669, "step": 273 }, { "epoch": 0.8968903436988543, "grad_norm": 0.2552024596786959, "learning_rate": 3.1028015099673957e-07, "loss": 2.4784, "step": 274 }, { "epoch": 0.900163666121113, "grad_norm": 0.2639761279668147, "learning_rate": 2.9077818180237693e-07, "loss": 2.4608, "step": 275 }, { "epoch": 0.9034369885433715, "grad_norm": 0.26158090716261023, "learning_rate": 2.7189079311587596e-07, "loss": 2.4317, "step": 276 }, { "epoch": 0.9067103109656302, "grad_norm": 0.26420232076269845, "learning_rate": 2.536204498499922e-07, "loss": 2.4715, "step": 277 }, { "epoch": 0.9099836333878887, "grad_norm": 0.2611776184759321, "learning_rate": 2.3596953638951093e-07, "loss": 2.4496, "step": 278 }, { "epoch": 0.9132569558101473, "grad_norm": 0.2566448882353072, "learning_rate": 2.1894035628006517e-07, "loss": 2.4535, "step": 279 }, { "epoch": 0.9165302782324058, "grad_norm": 0.2583889196721565, "learning_rate": 2.0253513192751374e-07, "loss": 2.4312, "step": 280 }, { "epoch": 0.9198036006546645, "grad_norm": 0.25005870928436097, "learning_rate": 1.867560043079031e-07, "loss": 2.4451, "step": 281 }, { "epoch": 0.9230769230769231, "grad_norm": 0.25955023725898285, "learning_rate": 1.7160503268806084e-07, "loss": 2.4591, "step": 282 }, { "epoch": 0.9263502454991817, "grad_norm": 0.25878369029655873, "learning_rate": 1.5708419435684463e-07, "loss": 2.4613, "step": 283 }, { "epoch": 0.9296235679214403, "grad_norm": 0.2559725798090891, "learning_rate": 1.4319538436709746e-07, "loss": 2.4422, "step": 284 }, { "epoch": 0.9328968903436988, "grad_norm": 0.2560397895273692, "learning_rate": 1.2994041528833267e-07, "loss": 2.4712, "step": 285 }, { "epoch": 0.9361702127659575, "grad_norm": 0.26063225688341923, "learning_rate": 1.1732101697018161e-07, "loss": 2.4452, "step": 286 }, { "epoch": 0.939443535188216, "grad_norm": 0.2582518091886246, "learning_rate": 1.0533883631663966e-07, "loss": 2.4559, "step": 287 }, { "epoch": 0.9427168576104746, "grad_norm": 0.26206927755347836, "learning_rate": 9.399543707113601e-08, "loss": 2.477, "step": 288 }, { "epoch": 0.9459901800327333, "grad_norm": 0.26042235837250066, "learning_rate": 8.329229961245355e-08, "loss": 2.4496, "step": 289 }, { "epoch": 0.9492635024549918, "grad_norm": 0.257835901510826, "learning_rate": 7.32308207615351e-08, "loss": 2.4546, "step": 290 }, { "epoch": 0.9525368248772504, "grad_norm": 0.26136373848402616, "learning_rate": 6.381231359918638e-08, "loss": 2.4491, "step": 291 }, { "epoch": 0.955810147299509, "grad_norm": 0.25188481875716506, "learning_rate": 5.503800729471376e-08, "loss": 2.4425, "step": 292 }, { "epoch": 0.9590834697217676, "grad_norm": 0.25523476009318996, "learning_rate": 4.690904694550913e-08, "loss": 2.4415, "step": 293 }, { "epoch": 0.9623567921440261, "grad_norm": 0.2558833671558322, "learning_rate": 3.9426493427611177e-08, "loss": 2.4591, "step": 294 }, { "epoch": 0.9656301145662848, "grad_norm": 0.2607553411892932, "learning_rate": 3.25913232572489e-08, "loss": 2.4868, "step": 295 }, { "epoch": 0.9689034369885434, "grad_norm": 0.262015003675821, "learning_rate": 2.640442846340796e-08, "loss": 2.4675, "step": 296 }, { "epoch": 0.972176759410802, "grad_norm": 0.2578562996530708, "learning_rate": 2.0866616471409974e-08, "loss": 2.4707, "step": 297 }, { "epoch": 0.9754500818330606, "grad_norm": 0.2567381919735001, "learning_rate": 1.5978609997542306e-08, "loss": 2.4582, "step": 298 }, { "epoch": 0.9787234042553191, "grad_norm": 0.2549822571991868, "learning_rate": 1.174104695473688e-08, "loss": 2.4825, "step": 299 }, { "epoch": 0.9819967266775778, "grad_norm": 0.2554549372989953, "learning_rate": 8.15448036932176e-09, "loss": 2.4719, "step": 300 }, { "epoch": 0.9852700490998363, "grad_norm": 0.263381774850731, "learning_rate": 5.219378308845558e-09, "loss": 2.4653, "step": 301 }, { "epoch": 0.9885433715220949, "grad_norm": 0.2528687928806352, "learning_rate": 2.9361238209935085e-09, "loss": 2.4689, "step": 302 }, { "epoch": 0.9918166939443536, "grad_norm": 0.2610005338150273, "learning_rate": 1.305014883595801e-09, "loss": 2.4555, "step": 303 }, { "epoch": 0.9950900163666121, "grad_norm": 0.26402197276145056, "learning_rate": 3.262643657425679e-10, "loss": 2.4546, "step": 304 }, { "epoch": 0.9983633387888707, "grad_norm": 0.25445050680135267, "learning_rate": 0.0, "loss": 2.4702, "step": 305 } ], "logging_steps": 1, "max_steps": 305, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.568140075322573e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }