{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9994739610731194, "eval_steps": 500, "global_step": 1425, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0007013852358407856, "grad_norm": NaN, "learning_rate": 0.0, "loss": 0.9295, "step": 1 }, { "epoch": 0.0014027704716815712, "grad_norm": NaN, "learning_rate": 0.0, "loss": 1.432, "step": 2 }, { "epoch": 0.0021041557075223566, "grad_norm": NaN, "learning_rate": 0.0, "loss": 0.3822, "step": 3 }, { "epoch": 0.0028055409433631424, "grad_norm": NaN, "learning_rate": 0.0, "loss": 0.6542, "step": 4 }, { "epoch": 0.003506926179203928, "grad_norm": 29.30962562561035, "learning_rate": 1.7543859649122808e-07, "loss": 0.3371, "step": 5 }, { "epoch": 0.004208311415044713, "grad_norm": 34.5477294921875, "learning_rate": 3.5087719298245616e-07, "loss": 0.1946, "step": 6 }, { "epoch": 0.004909696650885499, "grad_norm": 55.77326965332031, "learning_rate": 5.263157894736843e-07, "loss": 0.5682, "step": 7 }, { "epoch": 0.005611081886726285, "grad_norm": 52.84772491455078, "learning_rate": 7.017543859649123e-07, "loss": 0.3959, "step": 8 }, { "epoch": 0.00631246712256707, "grad_norm": 44.6428337097168, "learning_rate": 8.771929824561404e-07, "loss": 0.2899, "step": 9 }, { "epoch": 0.007013852358407856, "grad_norm": 69.89168548583984, "learning_rate": 1.0526315789473685e-06, "loss": 0.6255, "step": 10 }, { "epoch": 0.007715237594248641, "grad_norm": NaN, "learning_rate": 1.0526315789473685e-06, "loss": 0.4706, "step": 11 }, { "epoch": 0.008416622830089426, "grad_norm": 65.1708755493164, "learning_rate": 1.2280701754385965e-06, "loss": 0.6875, "step": 12 }, { "epoch": 0.009118008065930211, "grad_norm": 62.44850158691406, "learning_rate": 1.4035087719298246e-06, "loss": 0.8691, "step": 13 }, { "epoch": 0.009819393301770998, "grad_norm": 50.52750778198242, "learning_rate": 1.5789473684210528e-06, "loss": 0.5682, "step": 14 }, { "epoch": 0.010520778537611783, "grad_norm": 38.50554656982422, "learning_rate": 1.7543859649122807e-06, "loss": 0.4118, "step": 15 }, { "epoch": 0.01122216377345257, "grad_norm": Infinity, "learning_rate": 1.7543859649122807e-06, "loss": 0.4622, "step": 16 }, { "epoch": 0.011923549009293355, "grad_norm": 41.367347717285156, "learning_rate": 1.929824561403509e-06, "loss": 0.2331, "step": 17 }, { "epoch": 0.01262493424513414, "grad_norm": 47.666603088378906, "learning_rate": 2.105263157894737e-06, "loss": 0.489, "step": 18 }, { "epoch": 0.013326319480974926, "grad_norm": NaN, "learning_rate": 2.105263157894737e-06, "loss": 6.2172, "step": 19 }, { "epoch": 0.014027704716815711, "grad_norm": 161.6815643310547, "learning_rate": 2.2807017543859652e-06, "loss": 1.9947, "step": 20 }, { "epoch": 0.014729089952656496, "grad_norm": 14.031229019165039, "learning_rate": 2.456140350877193e-06, "loss": 0.0529, "step": 21 }, { "epoch": 0.015430475188497283, "grad_norm": 51.389862060546875, "learning_rate": 2.631578947368421e-06, "loss": 0.3272, "step": 22 }, { "epoch": 0.016131860424338066, "grad_norm": 16.31728744506836, "learning_rate": 2.8070175438596493e-06, "loss": 0.0668, "step": 23 }, { "epoch": 0.016833245660178853, "grad_norm": 32.19120407104492, "learning_rate": 2.9824561403508774e-06, "loss": 0.1298, "step": 24 }, { "epoch": 0.01753463089601964, "grad_norm": 8.127516746520996, "learning_rate": 3.1578947368421056e-06, "loss": 0.0245, "step": 25 }, { "epoch": 0.018236016131860423, "grad_norm": 31.48802375793457, "learning_rate": 3.3333333333333333e-06, "loss": 0.1808, "step": 26 }, { "epoch": 0.01893740136770121, "grad_norm": 174.7755584716797, "learning_rate": 3.5087719298245615e-06, "loss": 1.5976, "step": 27 }, { "epoch": 0.019638786603541996, "grad_norm": 173.22021484375, "learning_rate": 3.6842105263157892e-06, "loss": 2.0731, "step": 28 }, { "epoch": 0.020340171839382783, "grad_norm": 16.176433563232422, "learning_rate": 3.859649122807018e-06, "loss": 0.0533, "step": 29 }, { "epoch": 0.021041557075223566, "grad_norm": 25.456438064575195, "learning_rate": 4.035087719298246e-06, "loss": 0.2239, "step": 30 }, { "epoch": 0.021742942311064353, "grad_norm": 2.50178861618042, "learning_rate": 4.210526315789474e-06, "loss": 0.007, "step": 31 }, { "epoch": 0.02244432754690514, "grad_norm": 8.369728088378906, "learning_rate": 4.3859649122807014e-06, "loss": 0.0312, "step": 32 }, { "epoch": 0.023145712782745922, "grad_norm": 1.419299602508545, "learning_rate": 4.5614035087719304e-06, "loss": 0.0039, "step": 33 }, { "epoch": 0.02384709801858671, "grad_norm": 21.602149963378906, "learning_rate": 4.736842105263159e-06, "loss": 0.1907, "step": 34 }, { "epoch": 0.024548483254427496, "grad_norm": 96.2072525024414, "learning_rate": 4.912280701754386e-06, "loss": 0.4621, "step": 35 }, { "epoch": 0.02524986849026828, "grad_norm": 186.85726928710938, "learning_rate": 5.087719298245614e-06, "loss": 3.6381, "step": 36 }, { "epoch": 0.025951253726109066, "grad_norm": 222.0509033203125, "learning_rate": 5.263157894736842e-06, "loss": 2.1687, "step": 37 }, { "epoch": 0.026652638961949852, "grad_norm": 395.810791015625, "learning_rate": 5.43859649122807e-06, "loss": 4.1469, "step": 38 }, { "epoch": 0.027354024197790636, "grad_norm": 24.080060958862305, "learning_rate": 5.6140350877192985e-06, "loss": 0.0944, "step": 39 }, { "epoch": 0.028055409433631422, "grad_norm": 23.396860122680664, "learning_rate": 5.789473684210527e-06, "loss": 0.2287, "step": 40 }, { "epoch": 0.02875679466947221, "grad_norm": 26.664474487304688, "learning_rate": 5.964912280701755e-06, "loss": 0.2467, "step": 41 }, { "epoch": 0.029458179905312992, "grad_norm": 0.3562469482421875, "learning_rate": 6.140350877192982e-06, "loss": 0.0005, "step": 42 }, { "epoch": 0.03015956514115378, "grad_norm": Infinity, "learning_rate": 6.140350877192982e-06, "loss": 1.6967, "step": 43 }, { "epoch": 0.030860950376994566, "grad_norm": 0.1253829002380371, "learning_rate": 6.315789473684211e-06, "loss": 0.0003, "step": 44 }, { "epoch": 0.03156233561283535, "grad_norm": 18.994827270507812, "learning_rate": 6.4912280701754385e-06, "loss": 0.233, "step": 45 }, { "epoch": 0.03226372084867613, "grad_norm": 34.921852111816406, "learning_rate": 6.666666666666667e-06, "loss": 0.3528, "step": 46 }, { "epoch": 0.03296510608451692, "grad_norm": 202.69903564453125, "learning_rate": 6.842105263157896e-06, "loss": 2.7552, "step": 47 }, { "epoch": 0.033666491320357705, "grad_norm": 359.0797424316406, "learning_rate": 7.017543859649123e-06, "loss": 2.112, "step": 48 }, { "epoch": 0.03436787655619849, "grad_norm": 3.0262975692749023, "learning_rate": 7.192982456140351e-06, "loss": 0.0094, "step": 49 }, { "epoch": 0.03506926179203928, "grad_norm": 504.8307189941406, "learning_rate": 7.3684210526315784e-06, "loss": 6.4525, "step": 50 }, { "epoch": 0.03577064702788006, "grad_norm": 7.283820152282715, "learning_rate": 7.5438596491228074e-06, "loss": 0.0419, "step": 51 }, { "epoch": 0.036472032263720845, "grad_norm": 10.193113327026367, "learning_rate": 7.719298245614036e-06, "loss": 0.0805, "step": 52 }, { "epoch": 0.037173417499561635, "grad_norm": 21.247915267944336, "learning_rate": 7.894736842105263e-06, "loss": 0.1385, "step": 53 }, { "epoch": 0.03787480273540242, "grad_norm": 12.750287055969238, "learning_rate": 8.070175438596492e-06, "loss": 0.1786, "step": 54 }, { "epoch": 0.03857618797124321, "grad_norm": 186.3012237548828, "learning_rate": 8.245614035087721e-06, "loss": 2.5431, "step": 55 }, { "epoch": 0.03927757320708399, "grad_norm": 0.6861621141433716, "learning_rate": 8.421052631578948e-06, "loss": 0.0019, "step": 56 }, { "epoch": 0.039978958442924775, "grad_norm": 0.36288923025131226, "learning_rate": 8.596491228070176e-06, "loss": 0.0008, "step": 57 }, { "epoch": 0.040680343678765565, "grad_norm": 0.08618122339248657, "learning_rate": 8.771929824561403e-06, "loss": 0.0001, "step": 58 }, { "epoch": 0.04138172891460635, "grad_norm": 341.2626037597656, "learning_rate": 8.947368421052632e-06, "loss": 3.6742, "step": 59 }, { "epoch": 0.04208311415044713, "grad_norm": 5.118091583251953, "learning_rate": 9.122807017543861e-06, "loss": 0.0128, "step": 60 }, { "epoch": 0.04278449938628792, "grad_norm": 3.345242977142334, "learning_rate": 9.298245614035088e-06, "loss": 0.0095, "step": 61 }, { "epoch": 0.043485884622128705, "grad_norm": 319.7175598144531, "learning_rate": 9.473684210526317e-06, "loss": 4.3047, "step": 62 }, { "epoch": 0.04418726985796949, "grad_norm": 279.8736267089844, "learning_rate": 9.649122807017545e-06, "loss": 1.8524, "step": 63 }, { "epoch": 0.04488865509381028, "grad_norm": 187.50588989257812, "learning_rate": 9.824561403508772e-06, "loss": 2.2166, "step": 64 }, { "epoch": 0.04559004032965106, "grad_norm": 3.1667685508728027, "learning_rate": 1e-05, "loss": 0.0051, "step": 65 }, { "epoch": 0.046291425565491845, "grad_norm": 347.2233581542969, "learning_rate": 1.0175438596491228e-05, "loss": 6.7317, "step": 66 }, { "epoch": 0.046992810801332635, "grad_norm": 0.21984274685382843, "learning_rate": 1.0350877192982457e-05, "loss": 0.0006, "step": 67 }, { "epoch": 0.04769419603717342, "grad_norm": 657.2495727539062, "learning_rate": 1.0526315789473684e-05, "loss": 12.4963, "step": 68 }, { "epoch": 0.0483955812730142, "grad_norm": 180.615234375, "learning_rate": 1.0701754385964913e-05, "loss": 3.2259, "step": 69 }, { "epoch": 0.04909696650885499, "grad_norm": 0.4523675739765167, "learning_rate": 1.087719298245614e-05, "loss": 0.0011, "step": 70 }, { "epoch": 0.049798351744695775, "grad_norm": 241.3996124267578, "learning_rate": 1.1052631578947368e-05, "loss": 2.9636, "step": 71 }, { "epoch": 0.05049973698053656, "grad_norm": 0.21977363526821136, "learning_rate": 1.1228070175438597e-05, "loss": 0.0005, "step": 72 }, { "epoch": 0.05120112221637735, "grad_norm": 102.25936889648438, "learning_rate": 1.1403508771929824e-05, "loss": 0.4978, "step": 73 }, { "epoch": 0.05190250745221813, "grad_norm": 24.031539916992188, "learning_rate": 1.1578947368421053e-05, "loss": 0.119, "step": 74 }, { "epoch": 0.052603892688058915, "grad_norm": 187.41812133789062, "learning_rate": 1.1754385964912282e-05, "loss": 1.0223, "step": 75 }, { "epoch": 0.053305277923899705, "grad_norm": 45.128421783447266, "learning_rate": 1.192982456140351e-05, "loss": 0.4026, "step": 76 }, { "epoch": 0.05400666315974049, "grad_norm": 12.320211410522461, "learning_rate": 1.2105263157894737e-05, "loss": 0.1124, "step": 77 }, { "epoch": 0.05470804839558127, "grad_norm": 26.743837356567383, "learning_rate": 1.2280701754385964e-05, "loss": 0.1488, "step": 78 }, { "epoch": 0.05540943363142206, "grad_norm": 30.013471603393555, "learning_rate": 1.2456140350877193e-05, "loss": 0.3796, "step": 79 }, { "epoch": 0.056110818867262845, "grad_norm": 38.910221099853516, "learning_rate": 1.2631578947368422e-05, "loss": 0.7097, "step": 80 }, { "epoch": 0.05681220410310363, "grad_norm": 26.001049041748047, "learning_rate": 1.2807017543859651e-05, "loss": 0.1451, "step": 81 }, { "epoch": 0.05751358933894442, "grad_norm": 12.577742576599121, "learning_rate": 1.2982456140350877e-05, "loss": 0.0548, "step": 82 }, { "epoch": 0.0582149745747852, "grad_norm": 334.09478759765625, "learning_rate": 1.3157894736842106e-05, "loss": 8.8326, "step": 83 }, { "epoch": 0.058916359810625984, "grad_norm": 33.031837463378906, "learning_rate": 1.3333333333333333e-05, "loss": 0.2555, "step": 84 }, { "epoch": 0.059617745046466775, "grad_norm": 47.51133346557617, "learning_rate": 1.3508771929824562e-05, "loss": 0.4288, "step": 85 }, { "epoch": 0.06031913028230756, "grad_norm": 41.160430908203125, "learning_rate": 1.3684210526315791e-05, "loss": 0.3075, "step": 86 }, { "epoch": 0.06102051551814834, "grad_norm": 9.073790550231934, "learning_rate": 1.3859649122807017e-05, "loss": 0.045, "step": 87 }, { "epoch": 0.06172190075398913, "grad_norm": 352.0498352050781, "learning_rate": 1.4035087719298246e-05, "loss": 1.3396, "step": 88 }, { "epoch": 0.062423285989829914, "grad_norm": 36.5557746887207, "learning_rate": 1.4210526315789475e-05, "loss": 0.1926, "step": 89 }, { "epoch": 0.0631246712256707, "grad_norm": 6.991324424743652, "learning_rate": 1.4385964912280702e-05, "loss": 0.0341, "step": 90 }, { "epoch": 0.06382605646151149, "grad_norm": 6.0483078956604, "learning_rate": 1.4561403508771931e-05, "loss": 0.0219, "step": 91 }, { "epoch": 0.06452744169735226, "grad_norm": 4.510119438171387, "learning_rate": 1.4736842105263157e-05, "loss": 0.0161, "step": 92 }, { "epoch": 0.06522882693319305, "grad_norm": 0.8209487199783325, "learning_rate": 1.4912280701754386e-05, "loss": 0.0022, "step": 93 }, { "epoch": 0.06593021216903384, "grad_norm": 0.0308829378336668, "learning_rate": 1.5087719298245615e-05, "loss": 0.0001, "step": 94 }, { "epoch": 0.06663159740487462, "grad_norm": 171.68765258789062, "learning_rate": 1.5263157894736842e-05, "loss": 4.8668, "step": 95 }, { "epoch": 0.06733298264071541, "grad_norm": 0.0008907412411645055, "learning_rate": 1.543859649122807e-05, "loss": 0.0, "step": 96 }, { "epoch": 0.0680343678765562, "grad_norm": 0.0062671443447470665, "learning_rate": 1.56140350877193e-05, "loss": 0.0, "step": 97 }, { "epoch": 0.06873575311239698, "grad_norm": 0.02987469919025898, "learning_rate": 1.5789473684210526e-05, "loss": 0.0001, "step": 98 }, { "epoch": 0.06943713834823777, "grad_norm": 0.00033108692150563, "learning_rate": 1.5964912280701755e-05, "loss": 0.0, "step": 99 }, { "epoch": 0.07013852358407856, "grad_norm": 0.21173974871635437, "learning_rate": 1.6140350877192984e-05, "loss": 0.0003, "step": 100 }, { "epoch": 0.07083990881991933, "grad_norm": 0.0003154289734084159, "learning_rate": 1.6315789473684213e-05, "loss": 0.0, "step": 101 }, { "epoch": 0.07154129405576012, "grad_norm": 0.0001488685520598665, "learning_rate": 1.6491228070175442e-05, "loss": 0.0, "step": 102 }, { "epoch": 0.07224267929160091, "grad_norm": 1.3430032595351804e-05, "learning_rate": 1.6666666666666667e-05, "loss": 0.0, "step": 103 }, { "epoch": 0.07294406452744169, "grad_norm": 0.011110284365713596, "learning_rate": 1.6842105263157896e-05, "loss": 0.0, "step": 104 }, { "epoch": 0.07364544976328248, "grad_norm": 1.7930891772266477e-05, "learning_rate": 1.7017543859649125e-05, "loss": 0.0, "step": 105 }, { "epoch": 0.07434683499912327, "grad_norm": 5.653702146446449e-07, "learning_rate": 1.719298245614035e-05, "loss": 0.0, "step": 106 }, { "epoch": 0.07504822023496406, "grad_norm": 135.4245147705078, "learning_rate": 1.736842105263158e-05, "loss": 6.4792, "step": 107 }, { "epoch": 0.07574960547080484, "grad_norm": 153.99264526367188, "learning_rate": 1.7543859649122806e-05, "loss": 6.485, "step": 108 }, { "epoch": 0.07645099070664563, "grad_norm": 159.35337829589844, "learning_rate": 1.7719298245614035e-05, "loss": 2.4936, "step": 109 }, { "epoch": 0.07715237594248642, "grad_norm": 0.00017455461784265935, "learning_rate": 1.7894736842105264e-05, "loss": 0.0, "step": 110 }, { "epoch": 0.0778537611783272, "grad_norm": 146.1802978515625, "learning_rate": 1.8070175438596493e-05, "loss": 4.3088, "step": 111 }, { "epoch": 0.07855514641416798, "grad_norm": 0.022132212296128273, "learning_rate": 1.8245614035087722e-05, "loss": 0.0, "step": 112 }, { "epoch": 0.07925653165000877, "grad_norm": 160.43975830078125, "learning_rate": 1.8421052631578947e-05, "loss": 4.5829, "step": 113 }, { "epoch": 0.07995791688584955, "grad_norm": 164.45138549804688, "learning_rate": 1.8596491228070176e-05, "loss": 4.1681, "step": 114 }, { "epoch": 0.08065930212169034, "grad_norm": 0.02766432799398899, "learning_rate": 1.8771929824561405e-05, "loss": 0.0001, "step": 115 }, { "epoch": 0.08136068735753113, "grad_norm": 163.59400939941406, "learning_rate": 1.8947368421052634e-05, "loss": 3.0252, "step": 116 }, { "epoch": 0.08206207259337191, "grad_norm": 7.11077356338501, "learning_rate": 1.9122807017543863e-05, "loss": 0.0309, "step": 117 }, { "epoch": 0.0827634578292127, "grad_norm": 15.318146705627441, "learning_rate": 1.929824561403509e-05, "loss": 0.0856, "step": 118 }, { "epoch": 0.08346484306505349, "grad_norm": 9.1824951171875, "learning_rate": 1.9473684210526315e-05, "loss": 0.0416, "step": 119 }, { "epoch": 0.08416622830089426, "grad_norm": 22.720653533935547, "learning_rate": 1.9649122807017544e-05, "loss": 0.1365, "step": 120 }, { "epoch": 0.08486761353673505, "grad_norm": 26.200096130371094, "learning_rate": 1.9824561403508773e-05, "loss": 0.2579, "step": 121 }, { "epoch": 0.08556899877257584, "grad_norm": 15.874689102172852, "learning_rate": 2e-05, "loss": 0.0832, "step": 122 }, { "epoch": 0.08627038400841662, "grad_norm": 11.818103790283203, "learning_rate": 2.0175438596491227e-05, "loss": 0.0541, "step": 123 }, { "epoch": 0.08697176924425741, "grad_norm": 116.5608139038086, "learning_rate": 2.0350877192982456e-05, "loss": 0.9623, "step": 124 }, { "epoch": 0.0876731544800982, "grad_norm": 10.346467018127441, "learning_rate": 2.0526315789473685e-05, "loss": 0.0562, "step": 125 }, { "epoch": 0.08837453971593898, "grad_norm": 137.41136169433594, "learning_rate": 2.0701754385964914e-05, "loss": 1.1566, "step": 126 }, { "epoch": 0.08907592495177977, "grad_norm": 2.9712822437286377, "learning_rate": 2.0877192982456143e-05, "loss": 0.0104, "step": 127 }, { "epoch": 0.08977731018762056, "grad_norm": 18.723491668701172, "learning_rate": 2.105263157894737e-05, "loss": 0.0667, "step": 128 }, { "epoch": 0.09047869542346133, "grad_norm": 156.25631713867188, "learning_rate": 2.1228070175438598e-05, "loss": 2.4023, "step": 129 }, { "epoch": 0.09118008065930212, "grad_norm": 298.97552490234375, "learning_rate": 2.1403508771929827e-05, "loss": 6.4928, "step": 130 }, { "epoch": 0.09188146589514291, "grad_norm": 87.54296112060547, "learning_rate": 2.1578947368421053e-05, "loss": 0.5526, "step": 131 }, { "epoch": 0.09258285113098369, "grad_norm": 0.25955039262771606, "learning_rate": 2.175438596491228e-05, "loss": 0.0007, "step": 132 }, { "epoch": 0.09328423636682448, "grad_norm": 123.57398986816406, "learning_rate": 2.1929824561403507e-05, "loss": 1.0402, "step": 133 }, { "epoch": 0.09398562160266527, "grad_norm": 238.33229064941406, "learning_rate": 2.2105263157894736e-05, "loss": 2.4997, "step": 134 }, { "epoch": 0.09468700683850605, "grad_norm": 13.356133460998535, "learning_rate": 2.2280701754385965e-05, "loss": 0.0549, "step": 135 }, { "epoch": 0.09538839207434684, "grad_norm": 44.56560134887695, "learning_rate": 2.2456140350877194e-05, "loss": 0.4078, "step": 136 }, { "epoch": 0.09608977731018763, "grad_norm": 25.394866943359375, "learning_rate": 2.2631578947368423e-05, "loss": 0.1681, "step": 137 }, { "epoch": 0.0967911625460284, "grad_norm": 28.068824768066406, "learning_rate": 2.280701754385965e-05, "loss": 0.174, "step": 138 }, { "epoch": 0.0974925477818692, "grad_norm": 27.76630401611328, "learning_rate": 2.2982456140350878e-05, "loss": 0.1515, "step": 139 }, { "epoch": 0.09819393301770998, "grad_norm": 21.082311630249023, "learning_rate": 2.3157894736842107e-05, "loss": 0.1379, "step": 140 }, { "epoch": 0.09889531825355076, "grad_norm": 16.697647094726562, "learning_rate": 2.3333333333333336e-05, "loss": 0.1066, "step": 141 }, { "epoch": 0.09959670348939155, "grad_norm": 127.46953582763672, "learning_rate": 2.3508771929824565e-05, "loss": 1.3053, "step": 142 }, { "epoch": 0.10029808872523234, "grad_norm": 6.2219648361206055, "learning_rate": 2.368421052631579e-05, "loss": 0.0323, "step": 143 }, { "epoch": 0.10099947396107312, "grad_norm": 0.1309046596288681, "learning_rate": 2.385964912280702e-05, "loss": 0.0004, "step": 144 }, { "epoch": 0.1017008591969139, "grad_norm": 149.50645446777344, "learning_rate": 2.4035087719298245e-05, "loss": 1.8353, "step": 145 }, { "epoch": 0.1024022444327547, "grad_norm": 159.4170684814453, "learning_rate": 2.4210526315789474e-05, "loss": 3.2714, "step": 146 }, { "epoch": 0.10310362966859547, "grad_norm": 0.038805797696113586, "learning_rate": 2.4385964912280703e-05, "loss": 0.0001, "step": 147 }, { "epoch": 0.10380501490443626, "grad_norm": 301.7401123046875, "learning_rate": 2.456140350877193e-05, "loss": 5.5114, "step": 148 }, { "epoch": 0.10450640014027705, "grad_norm": 160.01580810546875, "learning_rate": 2.4736842105263158e-05, "loss": 3.0398, "step": 149 }, { "epoch": 0.10520778537611783, "grad_norm": 1.0593466758728027, "learning_rate": 2.4912280701754387e-05, "loss": 0.0026, "step": 150 }, { "epoch": 0.10590917061195862, "grad_norm": 0.1588684320449829, "learning_rate": 2.5087719298245616e-05, "loss": 0.0004, "step": 151 }, { "epoch": 0.10661055584779941, "grad_norm": 0.12831349670886993, "learning_rate": 2.5263157894736845e-05, "loss": 0.0004, "step": 152 }, { "epoch": 0.10731194108364019, "grad_norm": 1.908389925956726, "learning_rate": 2.5438596491228074e-05, "loss": 0.0062, "step": 153 }, { "epoch": 0.10801332631948098, "grad_norm": 0.9305810928344727, "learning_rate": 2.5614035087719303e-05, "loss": 0.0029, "step": 154 }, { "epoch": 0.10871471155532177, "grad_norm": 149.55746459960938, "learning_rate": 2.578947368421053e-05, "loss": 2.1251, "step": 155 }, { "epoch": 0.10941609679116254, "grad_norm": 83.3541030883789, "learning_rate": 2.5964912280701754e-05, "loss": 0.4841, "step": 156 }, { "epoch": 0.11011748202700333, "grad_norm": 0.37118402123451233, "learning_rate": 2.6140350877192983e-05, "loss": 0.0009, "step": 157 }, { "epoch": 0.11081886726284412, "grad_norm": 212.9443359375, "learning_rate": 2.6315789473684212e-05, "loss": 1.6639, "step": 158 }, { "epoch": 0.1115202524986849, "grad_norm": 1.386982798576355, "learning_rate": 2.6491228070175438e-05, "loss": 0.0049, "step": 159 }, { "epoch": 0.11222163773452569, "grad_norm": 151.23684692382812, "learning_rate": 2.6666666666666667e-05, "loss": 1.913, "step": 160 }, { "epoch": 0.11292302297036648, "grad_norm": 276.9901123046875, "learning_rate": 2.6842105263157896e-05, "loss": 5.0838, "step": 161 }, { "epoch": 0.11362440820620726, "grad_norm": 7.047855377197266, "learning_rate": 2.7017543859649125e-05, "loss": 0.0422, "step": 162 }, { "epoch": 0.11432579344204805, "grad_norm": 5.565479755401611, "learning_rate": 2.7192982456140354e-05, "loss": 0.0252, "step": 163 }, { "epoch": 0.11502717867788884, "grad_norm": 11.038246154785156, "learning_rate": 2.7368421052631583e-05, "loss": 0.064, "step": 164 }, { "epoch": 0.11572856391372961, "grad_norm": 1.4180843830108643, "learning_rate": 2.754385964912281e-05, "loss": 0.0048, "step": 165 }, { "epoch": 0.1164299491495704, "grad_norm": 131.0841064453125, "learning_rate": 2.7719298245614034e-05, "loss": 1.1797, "step": 166 }, { "epoch": 0.11713133438541119, "grad_norm": 286.4391784667969, "learning_rate": 2.7894736842105263e-05, "loss": 4.2848, "step": 167 }, { "epoch": 0.11783271962125197, "grad_norm": 0.354282945394516, "learning_rate": 2.8070175438596492e-05, "loss": 0.001, "step": 168 }, { "epoch": 0.11853410485709276, "grad_norm": 148.43812561035156, "learning_rate": 2.824561403508772e-05, "loss": 2.3637, "step": 169 }, { "epoch": 0.11923549009293355, "grad_norm": 3.4153621196746826, "learning_rate": 2.842105263157895e-05, "loss": 0.0121, "step": 170 }, { "epoch": 0.11993687532877433, "grad_norm": 6.812062740325928, "learning_rate": 2.8596491228070175e-05, "loss": 0.038, "step": 171 }, { "epoch": 0.12063826056461512, "grad_norm": 0.6381294131278992, "learning_rate": 2.8771929824561404e-05, "loss": 0.002, "step": 172 }, { "epoch": 0.1213396458004559, "grad_norm": 0.05846899002790451, "learning_rate": 2.8947368421052634e-05, "loss": 0.0002, "step": 173 }, { "epoch": 0.12204103103629668, "grad_norm": 140.1736297607422, "learning_rate": 2.9122807017543863e-05, "loss": 4.0868, "step": 174 }, { "epoch": 0.12274241627213747, "grad_norm": 1.7416075468063354, "learning_rate": 2.929824561403509e-05, "loss": 0.0057, "step": 175 }, { "epoch": 0.12344380150797826, "grad_norm": 0.35068169236183167, "learning_rate": 2.9473684210526314e-05, "loss": 0.0009, "step": 176 }, { "epoch": 0.12414518674381904, "grad_norm": 135.71665954589844, "learning_rate": 2.9649122807017543e-05, "loss": 2.1025, "step": 177 }, { "epoch": 0.12484657197965983, "grad_norm": 134.73190307617188, "learning_rate": 2.9824561403508772e-05, "loss": 1.8685, "step": 178 }, { "epoch": 0.1255479572155006, "grad_norm": 266.5934143066406, "learning_rate": 3e-05, "loss": 5.1122, "step": 179 }, { "epoch": 0.1262493424513414, "grad_norm": 383.29931640625, "learning_rate": 3.017543859649123e-05, "loss": 10.0944, "step": 180 }, { "epoch": 0.12695072768718219, "grad_norm": 3.7992992401123047, "learning_rate": 3.035087719298246e-05, "loss": 0.0161, "step": 181 }, { "epoch": 0.12765211292302298, "grad_norm": 206.0655975341797, "learning_rate": 3.0526315789473684e-05, "loss": 1.5303, "step": 182 }, { "epoch": 0.12835349815886377, "grad_norm": 13.431243896484375, "learning_rate": 3.0701754385964913e-05, "loss": 0.0912, "step": 183 }, { "epoch": 0.12905488339470453, "grad_norm": 16.69830894470215, "learning_rate": 3.087719298245614e-05, "loss": 0.0824, "step": 184 }, { "epoch": 0.12975626863054532, "grad_norm": 13.488637924194336, "learning_rate": 3.105263157894737e-05, "loss": 0.0746, "step": 185 }, { "epoch": 0.1304576538663861, "grad_norm": 10.065983772277832, "learning_rate": 3.12280701754386e-05, "loss": 0.0503, "step": 186 }, { "epoch": 0.1311590391022269, "grad_norm": 123.50023651123047, "learning_rate": 3.140350877192982e-05, "loss": 1.2874, "step": 187 }, { "epoch": 0.1318604243380677, "grad_norm": 1.7797540426254272, "learning_rate": 3.157894736842105e-05, "loss": 0.0067, "step": 188 }, { "epoch": 0.13256180957390848, "grad_norm": 0.33675551414489746, "learning_rate": 3.175438596491228e-05, "loss": 0.0011, "step": 189 }, { "epoch": 0.13326319480974924, "grad_norm": 118.4156723022461, "learning_rate": 3.192982456140351e-05, "loss": 1.2059, "step": 190 }, { "epoch": 0.13396458004559003, "grad_norm": 136.65228271484375, "learning_rate": 3.210526315789474e-05, "loss": 1.7856, "step": 191 }, { "epoch": 0.13466596528143082, "grad_norm": 145.7415313720703, "learning_rate": 3.228070175438597e-05, "loss": 2.8991, "step": 192 }, { "epoch": 0.1353673505172716, "grad_norm": 262.7060241699219, "learning_rate": 3.24561403508772e-05, "loss": 3.4817, "step": 193 }, { "epoch": 0.1360687357531124, "grad_norm": 0.25727707147598267, "learning_rate": 3.2631578947368426e-05, "loss": 0.0008, "step": 194 }, { "epoch": 0.1367701209889532, "grad_norm": 1.0702918767929077, "learning_rate": 3.2807017543859655e-05, "loss": 0.0037, "step": 195 }, { "epoch": 0.13747150622479395, "grad_norm": 1.5917607545852661, "learning_rate": 3.2982456140350884e-05, "loss": 0.0055, "step": 196 }, { "epoch": 0.13817289146063474, "grad_norm": 355.18511962890625, "learning_rate": 3.3157894736842106e-05, "loss": 11.4677, "step": 197 }, { "epoch": 0.13887427669647553, "grad_norm": 1.7947548627853394, "learning_rate": 3.3333333333333335e-05, "loss": 0.0074, "step": 198 }, { "epoch": 0.13957566193231633, "grad_norm": 1.5046889781951904, "learning_rate": 3.3508771929824564e-05, "loss": 0.0058, "step": 199 }, { "epoch": 0.14027704716815712, "grad_norm": 0.2553917467594147, "learning_rate": 3.368421052631579e-05, "loss": 0.0008, "step": 200 }, { "epoch": 0.1409784324039979, "grad_norm": 0.773563027381897, "learning_rate": 3.385964912280702e-05, "loss": 0.0027, "step": 201 }, { "epoch": 0.14167981763983867, "grad_norm": 166.71820068359375, "learning_rate": 3.403508771929825e-05, "loss": 6.1496, "step": 202 }, { "epoch": 0.14238120287567946, "grad_norm": 0.017217637971043587, "learning_rate": 3.421052631578947e-05, "loss": 0.0001, "step": 203 }, { "epoch": 0.14308258811152025, "grad_norm": 0.0036761912051588297, "learning_rate": 3.43859649122807e-05, "loss": 0.0, "step": 204 }, { "epoch": 0.14378397334736104, "grad_norm": 0.000679672637488693, "learning_rate": 3.456140350877193e-05, "loss": 0.0, "step": 205 }, { "epoch": 0.14448535858320183, "grad_norm": 0.00011312704009469599, "learning_rate": 3.473684210526316e-05, "loss": 0.0, "step": 206 }, { "epoch": 0.14518674381904262, "grad_norm": 0.00012077714200131595, "learning_rate": 3.491228070175438e-05, "loss": 0.0, "step": 207 }, { "epoch": 0.14588812905488338, "grad_norm": 0.00012628763215616345, "learning_rate": 3.508771929824561e-05, "loss": 0.0, "step": 208 }, { "epoch": 0.14658951429072417, "grad_norm": 1.7496255168225616e-05, "learning_rate": 3.526315789473684e-05, "loss": 0.0, "step": 209 }, { "epoch": 0.14729089952656496, "grad_norm": 1.5739196896902286e-05, "learning_rate": 3.543859649122807e-05, "loss": 0.0, "step": 210 }, { "epoch": 0.14799228476240575, "grad_norm": 0.08704076707363129, "learning_rate": 3.56140350877193e-05, "loss": 0.0001, "step": 211 }, { "epoch": 0.14869366999824654, "grad_norm": 0.0016435593133792281, "learning_rate": 3.578947368421053e-05, "loss": 0.0, "step": 212 }, { "epoch": 0.14939505523408733, "grad_norm": 1.5830031633377075, "learning_rate": 3.5964912280701756e-05, "loss": 0.005, "step": 213 }, { "epoch": 0.15009644046992812, "grad_norm": 0.00016859408060554415, "learning_rate": 3.6140350877192985e-05, "loss": 0.0, "step": 214 }, { "epoch": 0.15079782570576888, "grad_norm": 1.2645250535570085e-06, "learning_rate": 3.6315789473684214e-05, "loss": 0.0, "step": 215 }, { "epoch": 0.15149921094160967, "grad_norm": 336.5249328613281, "learning_rate": 3.6491228070175443e-05, "loss": 4.1939, "step": 216 }, { "epoch": 0.15220059617745046, "grad_norm": 9.010884127746976e-07, "learning_rate": 3.6666666666666666e-05, "loss": 0.0, "step": 217 }, { "epoch": 0.15290198141329125, "grad_norm": 5.530741304937692e-07, "learning_rate": 3.6842105263157895e-05, "loss": 0.0, "step": 218 }, { "epoch": 0.15360336664913204, "grad_norm": 222.7361297607422, "learning_rate": 3.7017543859649124e-05, "loss": 0.7712, "step": 219 }, { "epoch": 0.15430475188497284, "grad_norm": 123.51704406738281, "learning_rate": 3.719298245614035e-05, "loss": 7.8703, "step": 220 }, { "epoch": 0.1550061371208136, "grad_norm": 119.08999633789062, "learning_rate": 3.736842105263158e-05, "loss": 6.6427, "step": 221 }, { "epoch": 0.1557075223566544, "grad_norm": 0.005513960495591164, "learning_rate": 3.754385964912281e-05, "loss": 0.0, "step": 222 }, { "epoch": 0.15640890759249518, "grad_norm": 0.021047895774245262, "learning_rate": 3.771929824561404e-05, "loss": 0.0, "step": 223 }, { "epoch": 0.15711029282833597, "grad_norm": 0.001549599110148847, "learning_rate": 3.789473684210527e-05, "loss": 0.0, "step": 224 }, { "epoch": 0.15781167806417676, "grad_norm": 0.008736948482692242, "learning_rate": 3.80701754385965e-05, "loss": 0.0, "step": 225 }, { "epoch": 0.15851306330001755, "grad_norm": 0.03641926124691963, "learning_rate": 3.824561403508773e-05, "loss": 0.0001, "step": 226 }, { "epoch": 0.1592144485358583, "grad_norm": 0.026789208874106407, "learning_rate": 3.842105263157895e-05, "loss": 0.0001, "step": 227 }, { "epoch": 0.1599158337716991, "grad_norm": 130.79087829589844, "learning_rate": 3.859649122807018e-05, "loss": 3.4204, "step": 228 }, { "epoch": 0.1606172190075399, "grad_norm": 404.3518981933594, "learning_rate": 3.877192982456141e-05, "loss": 5.6741, "step": 229 }, { "epoch": 0.16131860424338068, "grad_norm": 5.39788818359375, "learning_rate": 3.894736842105263e-05, "loss": 0.0251, "step": 230 }, { "epoch": 0.16201998947922147, "grad_norm": 18.25702667236328, "learning_rate": 3.912280701754386e-05, "loss": 0.1059, "step": 231 }, { "epoch": 0.16272137471506226, "grad_norm": 2.5649735927581787, "learning_rate": 3.929824561403509e-05, "loss": 0.0108, "step": 232 }, { "epoch": 0.16342275995090302, "grad_norm": 389.410400390625, "learning_rate": 3.9473684210526316e-05, "loss": 4.1361, "step": 233 }, { "epoch": 0.16412414518674381, "grad_norm": 97.98786926269531, "learning_rate": 3.9649122807017545e-05, "loss": 1.0207, "step": 234 }, { "epoch": 0.1648255304225846, "grad_norm": 446.8147277832031, "learning_rate": 3.9824561403508774e-05, "loss": 2.3678, "step": 235 }, { "epoch": 0.1655269156584254, "grad_norm": 88.88761901855469, "learning_rate": 4e-05, "loss": 0.814, "step": 236 }, { "epoch": 0.16622830089426618, "grad_norm": 217.3586883544922, "learning_rate": 4.017543859649123e-05, "loss": 9.7291, "step": 237 }, { "epoch": 0.16692968613010697, "grad_norm": 106.2865219116211, "learning_rate": 4.0350877192982455e-05, "loss": 1.1716, "step": 238 }, { "epoch": 0.16763107136594774, "grad_norm": 154.2283477783203, "learning_rate": 4.0526315789473684e-05, "loss": 0.6351, "step": 239 }, { "epoch": 0.16833245660178853, "grad_norm": 37.964447021484375, "learning_rate": 4.070175438596491e-05, "loss": 0.3484, "step": 240 }, { "epoch": 0.16903384183762932, "grad_norm": 24.696714401245117, "learning_rate": 4.087719298245614e-05, "loss": 0.2762, "step": 241 }, { "epoch": 0.1697352270734701, "grad_norm": 82.12918853759766, "learning_rate": 4.105263157894737e-05, "loss": 0.9978, "step": 242 }, { "epoch": 0.1704366123093109, "grad_norm": 41.84835433959961, "learning_rate": 4.12280701754386e-05, "loss": 0.4799, "step": 243 }, { "epoch": 0.1711379975451517, "grad_norm": 45.34370803833008, "learning_rate": 4.140350877192983e-05, "loss": 0.4255, "step": 244 }, { "epoch": 0.17183938278099245, "grad_norm": 41.85076904296875, "learning_rate": 4.157894736842106e-05, "loss": 0.616, "step": 245 }, { "epoch": 0.17254076801683324, "grad_norm": 31.718475341796875, "learning_rate": 4.1754385964912287e-05, "loss": 0.2034, "step": 246 }, { "epoch": 0.17324215325267403, "grad_norm": 149.61940002441406, "learning_rate": 4.1929824561403516e-05, "loss": 1.5976, "step": 247 }, { "epoch": 0.17394353848851482, "grad_norm": 127.43466186523438, "learning_rate": 4.210526315789474e-05, "loss": 1.4465, "step": 248 }, { "epoch": 0.1746449237243556, "grad_norm": 36.885379791259766, "learning_rate": 4.228070175438597e-05, "loss": 0.5706, "step": 249 }, { "epoch": 0.1753463089601964, "grad_norm": 343.46649169921875, "learning_rate": 4.2456140350877196e-05, "loss": 5.0778, "step": 250 }, { "epoch": 0.17604769419603716, "grad_norm": 23.604751586914062, "learning_rate": 4.2631578947368425e-05, "loss": 0.1635, "step": 251 }, { "epoch": 0.17674907943187795, "grad_norm": 37.630043029785156, "learning_rate": 4.2807017543859654e-05, "loss": 0.5644, "step": 252 }, { "epoch": 0.17745046466771874, "grad_norm": 41.07538986206055, "learning_rate": 4.298245614035088e-05, "loss": 0.5337, "step": 253 }, { "epoch": 0.17815184990355953, "grad_norm": 66.12769317626953, "learning_rate": 4.3157894736842105e-05, "loss": 0.8491, "step": 254 }, { "epoch": 0.17885323513940032, "grad_norm": 38.53345489501953, "learning_rate": 4.3333333333333334e-05, "loss": 0.4, "step": 255 }, { "epoch": 0.17955462037524111, "grad_norm": 37.57467269897461, "learning_rate": 4.350877192982456e-05, "loss": 0.6272, "step": 256 }, { "epoch": 0.18025600561108188, "grad_norm": 27.241796493530273, "learning_rate": 4.368421052631579e-05, "loss": 0.2307, "step": 257 }, { "epoch": 0.18095739084692267, "grad_norm": 24.353778839111328, "learning_rate": 4.3859649122807014e-05, "loss": 0.1238, "step": 258 }, { "epoch": 0.18165877608276346, "grad_norm": 5.489215850830078, "learning_rate": 4.403508771929824e-05, "loss": 0.031, "step": 259 }, { "epoch": 0.18236016131860425, "grad_norm": 16.305925369262695, "learning_rate": 4.421052631578947e-05, "loss": 0.0398, "step": 260 }, { "epoch": 0.18306154655444504, "grad_norm": 6.403785228729248, "learning_rate": 4.43859649122807e-05, "loss": 0.0304, "step": 261 }, { "epoch": 0.18376293179028583, "grad_norm": 4.471805572509766, "learning_rate": 4.456140350877193e-05, "loss": 0.0077, "step": 262 }, { "epoch": 0.1844643170261266, "grad_norm": 0.9698334336280823, "learning_rate": 4.473684210526316e-05, "loss": 0.0031, "step": 263 }, { "epoch": 0.18516570226196738, "grad_norm": 0.014244407415390015, "learning_rate": 4.491228070175439e-05, "loss": 0.0, "step": 264 }, { "epoch": 0.18586708749780817, "grad_norm": 121.71415710449219, "learning_rate": 4.508771929824562e-05, "loss": 4.2289, "step": 265 }, { "epoch": 0.18656847273364896, "grad_norm": 0.02348833531141281, "learning_rate": 4.5263157894736846e-05, "loss": 0.0001, "step": 266 }, { "epoch": 0.18726985796948975, "grad_norm": 0.0016702099237591028, "learning_rate": 4.5438596491228075e-05, "loss": 0.0, "step": 267 }, { "epoch": 0.18797124320533054, "grad_norm": 0.0027486609760671854, "learning_rate": 4.56140350877193e-05, "loss": 0.0, "step": 268 }, { "epoch": 0.1886726284411713, "grad_norm": 0.0012574723223224282, "learning_rate": 4.5789473684210527e-05, "loss": 0.0, "step": 269 }, { "epoch": 0.1893740136770121, "grad_norm": 0.002944325562566519, "learning_rate": 4.5964912280701756e-05, "loss": 0.0, "step": 270 }, { "epoch": 0.19007539891285288, "grad_norm": 0.0020754581782966852, "learning_rate": 4.6140350877192985e-05, "loss": 0.0, "step": 271 }, { "epoch": 0.19077678414869367, "grad_norm": 0.008111892268061638, "learning_rate": 4.6315789473684214e-05, "loss": 0.0, "step": 272 }, { "epoch": 0.19147816938453446, "grad_norm": 118.25870513916016, "learning_rate": 4.649122807017544e-05, "loss": 3.5798, "step": 273 }, { "epoch": 0.19217955462037525, "grad_norm": 118.91422271728516, "learning_rate": 4.666666666666667e-05, "loss": 2.0933, "step": 274 }, { "epoch": 0.19288093985621602, "grad_norm": 0.1281914860010147, "learning_rate": 4.68421052631579e-05, "loss": 0.0004, "step": 275 }, { "epoch": 0.1935823250920568, "grad_norm": 0.9166204929351807, "learning_rate": 4.701754385964913e-05, "loss": 0.0034, "step": 276 }, { "epoch": 0.1942837103278976, "grad_norm": 3.7037837505340576, "learning_rate": 4.719298245614036e-05, "loss": 0.0179, "step": 277 }, { "epoch": 0.1949850955637384, "grad_norm": 0.716585099697113, "learning_rate": 4.736842105263158e-05, "loss": 0.0028, "step": 278 }, { "epoch": 0.19568648079957918, "grad_norm": 119.46430969238281, "learning_rate": 4.754385964912281e-05, "loss": 1.7075, "step": 279 }, { "epoch": 0.19638786603541997, "grad_norm": 0.07794260233640671, "learning_rate": 4.771929824561404e-05, "loss": 0.0003, "step": 280 }, { "epoch": 0.19708925127126073, "grad_norm": 119.65841674804688, "learning_rate": 4.789473684210526e-05, "loss": 3.3725, "step": 281 }, { "epoch": 0.19779063650710152, "grad_norm": 0.06007947400212288, "learning_rate": 4.807017543859649e-05, "loss": 0.0002, "step": 282 }, { "epoch": 0.1984920217429423, "grad_norm": 0.33029523491859436, "learning_rate": 4.824561403508772e-05, "loss": 0.0013, "step": 283 }, { "epoch": 0.1991934069787831, "grad_norm": 0.19337859749794006, "learning_rate": 4.842105263157895e-05, "loss": 0.0007, "step": 284 }, { "epoch": 0.1998947922146239, "grad_norm": 0.4252503514289856, "learning_rate": 4.859649122807018e-05, "loss": 0.0017, "step": 285 }, { "epoch": 0.20059617745046468, "grad_norm": 0.07317744940519333, "learning_rate": 4.8771929824561406e-05, "loss": 0.0003, "step": 286 }, { "epoch": 0.20129756268630544, "grad_norm": 0.01806594245135784, "learning_rate": 4.8947368421052635e-05, "loss": 0.0001, "step": 287 }, { "epoch": 0.20199894792214623, "grad_norm": 106.39412689208984, "learning_rate": 4.912280701754386e-05, "loss": 3.9844, "step": 288 }, { "epoch": 0.20270033315798702, "grad_norm": 0.008989217691123486, "learning_rate": 4.9298245614035086e-05, "loss": 0.0, "step": 289 }, { "epoch": 0.2034017183938278, "grad_norm": 0.05597059056162834, "learning_rate": 4.9473684210526315e-05, "loss": 0.0002, "step": 290 }, { "epoch": 0.2041031036296686, "grad_norm": 113.05389404296875, "learning_rate": 4.9649122807017544e-05, "loss": 2.5036, "step": 291 }, { "epoch": 0.2048044888655094, "grad_norm": 0.39913010597229004, "learning_rate": 4.9824561403508773e-05, "loss": 0.0015, "step": 292 }, { "epoch": 0.20550587410135016, "grad_norm": 0.944106936454773, "learning_rate": 5e-05, "loss": 0.0038, "step": 293 }, { "epoch": 0.20620725933719095, "grad_norm": 108.76079559326172, "learning_rate": 4.9999995795715716e-05, "loss": 1.4773, "step": 294 }, { "epoch": 0.20690864457303174, "grad_norm": 1.2869099378585815, "learning_rate": 4.999998318286425e-05, "loss": 0.0054, "step": 295 }, { "epoch": 0.20761002980887253, "grad_norm": 0.7701173424720764, "learning_rate": 4.999996216144987e-05, "loss": 0.0029, "step": 296 }, { "epoch": 0.20831141504471332, "grad_norm": 104.06690216064453, "learning_rate": 4.9999932731479625e-05, "loss": 1.2931, "step": 297 }, { "epoch": 0.2090128002805541, "grad_norm": 234.58531188964844, "learning_rate": 4.999989489296344e-05, "loss": 4.5688, "step": 298 }, { "epoch": 0.20971418551639487, "grad_norm": 108.04949188232422, "learning_rate": 4.999984864591401e-05, "loss": 1.5544, "step": 299 }, { "epoch": 0.21041557075223566, "grad_norm": 1.8640260696411133, "learning_rate": 4.999979399034691e-05, "loss": 0.0078, "step": 300 }, { "epoch": 0.21111695598807645, "grad_norm": 3.089526414871216, "learning_rate": 4.999973092628052e-05, "loss": 0.0138, "step": 301 }, { "epoch": 0.21181834122391724, "grad_norm": 1.4183728694915771, "learning_rate": 4.999965945373605e-05, "loss": 0.006, "step": 302 }, { "epoch": 0.21251972645975803, "grad_norm": 0.9733534455299377, "learning_rate": 4.9999579572737533e-05, "loss": 0.0038, "step": 303 }, { "epoch": 0.21322111169559882, "grad_norm": 121.11236572265625, "learning_rate": 4.9999491283311836e-05, "loss": 2.2543, "step": 304 }, { "epoch": 0.21392249693143958, "grad_norm": 117.718017578125, "learning_rate": 4.999939458548868e-05, "loss": 2.5007, "step": 305 }, { "epoch": 0.21462388216728037, "grad_norm": 0.013796065002679825, "learning_rate": 4.9999289479300557e-05, "loss": 0.0, "step": 306 }, { "epoch": 0.21532526740312116, "grad_norm": 116.46581268310547, "learning_rate": 4.999917596478283e-05, "loss": 2.6005, "step": 307 }, { "epoch": 0.21602665263896195, "grad_norm": 0.3713426887989044, "learning_rate": 4.999905404197368e-05, "loss": 0.0014, "step": 308 }, { "epoch": 0.21672803787480274, "grad_norm": 0.44403278827667236, "learning_rate": 4.999892371091411e-05, "loss": 0.0017, "step": 309 }, { "epoch": 0.21742942311064353, "grad_norm": 139.35816955566406, "learning_rate": 4.999878497164797e-05, "loss": 11.1347, "step": 310 }, { "epoch": 0.2181308083464843, "grad_norm": 0.20512554049491882, "learning_rate": 4.999863782422191e-05, "loss": 0.0007, "step": 311 }, { "epoch": 0.21883219358232509, "grad_norm": 88.1954345703125, "learning_rate": 4.9998482268685434e-05, "loss": 0.843, "step": 312 }, { "epoch": 0.21953357881816588, "grad_norm": 93.83118438720703, "learning_rate": 4.999831830509084e-05, "loss": 1.0804, "step": 313 }, { "epoch": 0.22023496405400667, "grad_norm": 5.042333602905273, "learning_rate": 4.99981459334933e-05, "loss": 0.0256, "step": 314 }, { "epoch": 0.22093634928984746, "grad_norm": 7.477559566497803, "learning_rate": 4.9997965153950775e-05, "loss": 0.0402, "step": 315 }, { "epoch": 0.22163773452568825, "grad_norm": 156.78961181640625, "learning_rate": 4.999777596652408e-05, "loss": 6.0824, "step": 316 }, { "epoch": 0.222339119761529, "grad_norm": 9.518619537353516, "learning_rate": 4.999757837127683e-05, "loss": 0.0577, "step": 317 }, { "epoch": 0.2230405049973698, "grad_norm": 2.860729455947876, "learning_rate": 4.9997372368275506e-05, "loss": 0.0135, "step": 318 }, { "epoch": 0.2237418902332106, "grad_norm": 1.1225600242614746, "learning_rate": 4.999715795758938e-05, "loss": 0.0049, "step": 319 }, { "epoch": 0.22444327546905138, "grad_norm": 0.12374955415725708, "learning_rate": 4.9996935139290575e-05, "loss": 0.0004, "step": 320 }, { "epoch": 0.22514466070489217, "grad_norm": 112.18521881103516, "learning_rate": 4.999670391345403e-05, "loss": 3.4372, "step": 321 }, { "epoch": 0.22584604594073296, "grad_norm": 0.0004930093418806791, "learning_rate": 4.999646428015752e-05, "loss": 0.0, "step": 322 }, { "epoch": 0.22654743117657372, "grad_norm": 0.00027333476464264095, "learning_rate": 4.9996216239481643e-05, "loss": 0.0, "step": 323 }, { "epoch": 0.2272488164124145, "grad_norm": 0.0001735862751957029, "learning_rate": 4.999595979150982e-05, "loss": 0.0, "step": 324 }, { "epoch": 0.2279502016482553, "grad_norm": 2.878743180190213e-05, "learning_rate": 4.9995694936328316e-05, "loss": 0.0, "step": 325 }, { "epoch": 0.2286515868840961, "grad_norm": 109.58158111572266, "learning_rate": 4.99954216740262e-05, "loss": 4.991, "step": 326 }, { "epoch": 0.22935297211993688, "grad_norm": 113.2159194946289, "learning_rate": 4.9995140004695396e-05, "loss": 4.7235, "step": 327 }, { "epoch": 0.23005435735577767, "grad_norm": 0.00015405479643959552, "learning_rate": 4.999484992843063e-05, "loss": 0.0, "step": 328 }, { "epoch": 0.23075574259161843, "grad_norm": 0.011504331603646278, "learning_rate": 4.999455144532947e-05, "loss": 0.0, "step": 329 }, { "epoch": 0.23145712782745922, "grad_norm": 0.001587590086273849, "learning_rate": 4.9994244555492315e-05, "loss": 0.0, "step": 330 }, { "epoch": 0.23215851306330001, "grad_norm": 0.0038678112905472517, "learning_rate": 4.999392925902238e-05, "loss": 0.0, "step": 331 }, { "epoch": 0.2328598982991408, "grad_norm": 0.044445931911468506, "learning_rate": 4.999360555602571e-05, "loss": 0.0001, "step": 332 }, { "epoch": 0.2335612835349816, "grad_norm": 108.20228576660156, "learning_rate": 4.999327344661118e-05, "loss": 2.0004, "step": 333 }, { "epoch": 0.23426266877082239, "grad_norm": 0.17047792673110962, "learning_rate": 4.9992932930890505e-05, "loss": 0.0005, "step": 334 }, { "epoch": 0.23496405400666315, "grad_norm": 0.39230290055274963, "learning_rate": 4.999258400897819e-05, "loss": 0.0016, "step": 335 }, { "epoch": 0.23566543924250394, "grad_norm": 191.8765106201172, "learning_rate": 4.999222668099163e-05, "loss": 2.4042, "step": 336 }, { "epoch": 0.23636682447834473, "grad_norm": 87.04983520507812, "learning_rate": 4.999186094705097e-05, "loss": 0.9342, "step": 337 }, { "epoch": 0.23706820971418552, "grad_norm": 14.748551368713379, "learning_rate": 4.999148680727925e-05, "loss": 0.1165, "step": 338 }, { "epoch": 0.2377695949500263, "grad_norm": 16.103792190551758, "learning_rate": 4.999110426180229e-05, "loss": 0.11, "step": 339 }, { "epoch": 0.2384709801858671, "grad_norm": 4.507053375244141, "learning_rate": 4.999071331074877e-05, "loss": 0.1421, "step": 340 }, { "epoch": 0.23917236542170786, "grad_norm": 8.822351455688477, "learning_rate": 4.999031395425018e-05, "loss": 0.2372, "step": 341 }, { "epoch": 0.23987375065754865, "grad_norm": 19.92313575744629, "learning_rate": 4.998990619244084e-05, "loss": 0.1537, "step": 342 }, { "epoch": 0.24057513589338944, "grad_norm": 12.00966739654541, "learning_rate": 4.998949002545789e-05, "loss": 0.0872, "step": 343 }, { "epoch": 0.24127652112923023, "grad_norm": 4.180269718170166, "learning_rate": 4.998906545344132e-05, "loss": 0.0218, "step": 344 }, { "epoch": 0.24197790636507102, "grad_norm": 0.596821129322052, "learning_rate": 4.998863247653392e-05, "loss": 0.0025, "step": 345 }, { "epoch": 0.2426792916009118, "grad_norm": 0.4522903263568878, "learning_rate": 4.998819109488132e-05, "loss": 0.0013, "step": 346 }, { "epoch": 0.24338067683675257, "grad_norm": 0.0007598726078867912, "learning_rate": 4.998774130863199e-05, "loss": 0.0, "step": 347 }, { "epoch": 0.24408206207259336, "grad_norm": 8.103512664092705e-05, "learning_rate": 4.998728311793719e-05, "loss": 0.0, "step": 348 }, { "epoch": 0.24478344730843415, "grad_norm": 105.80138397216797, "learning_rate": 4.998681652295104e-05, "loss": 4.8223, "step": 349 }, { "epoch": 0.24548483254427494, "grad_norm": 127.92369079589844, "learning_rate": 4.998634152383047e-05, "loss": 7.0698, "step": 350 }, { "epoch": 0.24618621778011573, "grad_norm": 2.636051704030251e-06, "learning_rate": 4.998585812073525e-05, "loss": 0.0, "step": 351 }, { "epoch": 0.24688760301595652, "grad_norm": 2.710710077735712e-06, "learning_rate": 4.9985366313827975e-05, "loss": 0.0, "step": 352 }, { "epoch": 0.2475889882517973, "grad_norm": 103.96717834472656, "learning_rate": 4.998486610327405e-05, "loss": 6.485, "step": 353 }, { "epoch": 0.24829037348763808, "grad_norm": 103.35551452636719, "learning_rate": 4.9984357489241715e-05, "loss": 5.7037, "step": 354 }, { "epoch": 0.24899175872347887, "grad_norm": 100.56452178955078, "learning_rate": 4.998384047190204e-05, "loss": 4.1908, "step": 355 }, { "epoch": 0.24969314395931966, "grad_norm": 200.88751220703125, "learning_rate": 4.998331505142893e-05, "loss": 4.8984, "step": 356 }, { "epoch": 0.2503945291951604, "grad_norm": 77.69247436523438, "learning_rate": 4.99827812279991e-05, "loss": 0.9449, "step": 357 }, { "epoch": 0.2510959144310012, "grad_norm": 26.622453689575195, "learning_rate": 4.9982239001792095e-05, "loss": 0.3613, "step": 358 }, { "epoch": 0.251797299666842, "grad_norm": 26.442935943603516, "learning_rate": 4.9981688372990284e-05, "loss": 0.5424, "step": 359 }, { "epoch": 0.2524986849026828, "grad_norm": 39.277618408203125, "learning_rate": 4.9981129341778886e-05, "loss": 0.9456, "step": 360 }, { "epoch": 0.2532000701385236, "grad_norm": 43.96021270751953, "learning_rate": 4.9980561908345916e-05, "loss": 0.6765, "step": 361 }, { "epoch": 0.25390145537436437, "grad_norm": 14.254291534423828, "learning_rate": 4.997998607288222e-05, "loss": 0.1276, "step": 362 }, { "epoch": 0.25460284061020516, "grad_norm": 0.6544491052627563, "learning_rate": 4.9979401835581476e-05, "loss": 0.0032, "step": 363 }, { "epoch": 0.25530422584604595, "grad_norm": 0.011080354452133179, "learning_rate": 4.99788091966402e-05, "loss": 0.0, "step": 364 }, { "epoch": 0.25600561108188674, "grad_norm": 284.8055114746094, "learning_rate": 4.997820815625771e-05, "loss": 7.6058, "step": 365 }, { "epoch": 0.25670699631772753, "grad_norm": 89.02140045166016, "learning_rate": 4.997759871463618e-05, "loss": 5.079, "step": 366 }, { "epoch": 0.2574083815535683, "grad_norm": 4.7222201828844845e-05, "learning_rate": 4.997698087198056e-05, "loss": 0.0, "step": 367 }, { "epoch": 0.25810976678940906, "grad_norm": 85.66144561767578, "learning_rate": 4.997635462849869e-05, "loss": 4.8183, "step": 368 }, { "epoch": 0.25881115202524985, "grad_norm": 0.0015366720035672188, "learning_rate": 4.997571998440118e-05, "loss": 0.0, "step": 369 }, { "epoch": 0.25951253726109064, "grad_norm": 100.14698028564453, "learning_rate": 4.99750769399015e-05, "loss": 2.2663, "step": 370 }, { "epoch": 0.2602139224969314, "grad_norm": 0.4829590618610382, "learning_rate": 4.997442549521592e-05, "loss": 0.0021, "step": 371 }, { "epoch": 0.2609153077327722, "grad_norm": 70.44209289550781, "learning_rate": 4.9973765650563564e-05, "loss": 0.7532, "step": 372 }, { "epoch": 0.261616692968613, "grad_norm": 272.1163024902344, "learning_rate": 4.9973097406166355e-05, "loss": 3.9935, "step": 373 }, { "epoch": 0.2623180782044538, "grad_norm": 25.12706184387207, "learning_rate": 4.997242076224906e-05, "loss": 0.3738, "step": 374 }, { "epoch": 0.2630194634402946, "grad_norm": 35.15785598754883, "learning_rate": 4.997173571903926e-05, "loss": 0.6808, "step": 375 }, { "epoch": 0.2637208486761354, "grad_norm": 27.46406364440918, "learning_rate": 4.9971042276767366e-05, "loss": 0.2568, "step": 376 }, { "epoch": 0.26442223391197617, "grad_norm": 6.682714462280273, "learning_rate": 4.997034043566661e-05, "loss": 0.0374, "step": 377 }, { "epoch": 0.26512361914781696, "grad_norm": 0.057287223637104034, "learning_rate": 4.9969630195973045e-05, "loss": 0.0001, "step": 378 }, { "epoch": 0.26582500438365775, "grad_norm": 0.006871811114251614, "learning_rate": 4.9968911557925564e-05, "loss": 0.0, "step": 379 }, { "epoch": 0.2665263896194985, "grad_norm": 0.0037054684944450855, "learning_rate": 4.996818452176587e-05, "loss": 0.0, "step": 380 }, { "epoch": 0.2672277748553393, "grad_norm": 123.15022277832031, "learning_rate": 4.9967449087738505e-05, "loss": 5.8346, "step": 381 }, { "epoch": 0.26792916009118006, "grad_norm": 121.81306457519531, "learning_rate": 4.996670525609082e-05, "loss": 4.3778, "step": 382 }, { "epoch": 0.26863054532702085, "grad_norm": 0.00026572938077151775, "learning_rate": 4.9965953027072996e-05, "loss": 0.0, "step": 383 }, { "epoch": 0.26933193056286164, "grad_norm": 198.09776306152344, "learning_rate": 4.996519240093806e-05, "loss": 5.2574, "step": 384 }, { "epoch": 0.27003331579870243, "grad_norm": 130.21214294433594, "learning_rate": 4.996442337794181e-05, "loss": 3.217, "step": 385 }, { "epoch": 0.2707347010345432, "grad_norm": 133.94178771972656, "learning_rate": 4.996364595834291e-05, "loss": 1.3412, "step": 386 }, { "epoch": 0.271436086270384, "grad_norm": 12.223856925964355, "learning_rate": 4.9962860142402853e-05, "loss": 0.0924, "step": 387 }, { "epoch": 0.2721374715062248, "grad_norm": 21.94683074951172, "learning_rate": 4.9962065930385934e-05, "loss": 0.1824, "step": 388 }, { "epoch": 0.2728388567420656, "grad_norm": 46.802406311035156, "learning_rate": 4.996126332255929e-05, "loss": 0.3585, "step": 389 }, { "epoch": 0.2735402419779064, "grad_norm": 7.21342134475708, "learning_rate": 4.9960452319192866e-05, "loss": 0.1045, "step": 390 }, { "epoch": 0.2742416272137472, "grad_norm": 33.5169677734375, "learning_rate": 4.995963292055943e-05, "loss": 0.248, "step": 391 }, { "epoch": 0.2749430124495879, "grad_norm": 16.911136627197266, "learning_rate": 4.995880512693458e-05, "loss": 0.0876, "step": 392 }, { "epoch": 0.2756443976854287, "grad_norm": 206.7276153564453, "learning_rate": 4.995796893859675e-05, "loss": 1.6484, "step": 393 }, { "epoch": 0.2763457829212695, "grad_norm": 94.29922485351562, "learning_rate": 4.9957124355827186e-05, "loss": 0.8745, "step": 394 }, { "epoch": 0.2770471681571103, "grad_norm": 156.6600799560547, "learning_rate": 4.9956271378909945e-05, "loss": 0.9049, "step": 395 }, { "epoch": 0.27774855339295107, "grad_norm": 6.1268744468688965, "learning_rate": 4.995541000813193e-05, "loss": 0.0343, "step": 396 }, { "epoch": 0.27844993862879186, "grad_norm": 1.731136679649353, "learning_rate": 4.995454024378285e-05, "loss": 0.0075, "step": 397 }, { "epoch": 0.27915132386463265, "grad_norm": 144.5174560546875, "learning_rate": 4.995366208615526e-05, "loss": 2.6072, "step": 398 }, { "epoch": 0.27985270910047344, "grad_norm": 0.19711287319660187, "learning_rate": 4.9952775535544494e-05, "loss": 0.0007, "step": 399 }, { "epoch": 0.28055409433631423, "grad_norm": 0.13065628707408905, "learning_rate": 4.9951880592248754e-05, "loss": 0.0004, "step": 400 }, { "epoch": 0.281255479572155, "grad_norm": 125.28437042236328, "learning_rate": 4.995097725656905e-05, "loss": 1.8792, "step": 401 }, { "epoch": 0.2819568648079958, "grad_norm": 111.38099670410156, "learning_rate": 4.9950065528809214e-05, "loss": 2.548, "step": 402 }, { "epoch": 0.2826582500438366, "grad_norm": 1.2473199367523193, "learning_rate": 4.994914540927589e-05, "loss": 0.0042, "step": 403 }, { "epoch": 0.28335963527967734, "grad_norm": 0.7474699020385742, "learning_rate": 4.994821689827856e-05, "loss": 0.0026, "step": 404 }, { "epoch": 0.2840610205155181, "grad_norm": 0.14196322858333588, "learning_rate": 4.99472799961295e-05, "loss": 0.0005, "step": 405 }, { "epoch": 0.2847624057513589, "grad_norm": 0.6420115828514099, "learning_rate": 4.994633470314387e-05, "loss": 0.0026, "step": 406 }, { "epoch": 0.2854637909871997, "grad_norm": 213.8779296875, "learning_rate": 4.994538101963958e-05, "loss": 2.1989, "step": 407 }, { "epoch": 0.2861651762230405, "grad_norm": 108.08120727539062, "learning_rate": 4.994441894593741e-05, "loss": 3.9965, "step": 408 }, { "epoch": 0.2868665614588813, "grad_norm": 0.40015965700149536, "learning_rate": 4.994344848236094e-05, "loss": 0.0016, "step": 409 }, { "epoch": 0.2875679466947221, "grad_norm": 125.80431365966797, "learning_rate": 4.9942469629236585e-05, "loss": 2.3139, "step": 410 }, { "epoch": 0.28826933193056287, "grad_norm": 101.27831268310547, "learning_rate": 4.994148238689357e-05, "loss": 2.1065, "step": 411 }, { "epoch": 0.28897071716640366, "grad_norm": 3.526338577270508, "learning_rate": 4.9940486755663946e-05, "loss": 0.015, "step": 412 }, { "epoch": 0.28967210240224445, "grad_norm": 2.501554489135742, "learning_rate": 4.993948273588259e-05, "loss": 0.0123, "step": 413 }, { "epoch": 0.29037348763808524, "grad_norm": 75.39434814453125, "learning_rate": 4.993847032788719e-05, "loss": 0.7107, "step": 414 }, { "epoch": 0.291074872873926, "grad_norm": 3.751978635787964, "learning_rate": 4.993744953201827e-05, "loss": 0.0183, "step": 415 }, { "epoch": 0.29177625810976676, "grad_norm": 5.31666374206543, "learning_rate": 4.9936420348619164e-05, "loss": 0.0385, "step": 416 }, { "epoch": 0.29247764334560755, "grad_norm": 104.77239227294922, "learning_rate": 4.9935382778036036e-05, "loss": 1.7353, "step": 417 }, { "epoch": 0.29317902858144834, "grad_norm": 92.44829559326172, "learning_rate": 4.993433682061786e-05, "loss": 1.1895, "step": 418 }, { "epoch": 0.29388041381728913, "grad_norm": 89.00714874267578, "learning_rate": 4.993328247671643e-05, "loss": 1.2323, "step": 419 }, { "epoch": 0.2945817990531299, "grad_norm": 172.867431640625, "learning_rate": 4.993221974668637e-05, "loss": 1.7303, "step": 420 }, { "epoch": 0.2952831842889707, "grad_norm": 19.69775390625, "learning_rate": 4.9931148630885127e-05, "loss": 0.1678, "step": 421 }, { "epoch": 0.2959845695248115, "grad_norm": 22.828763961791992, "learning_rate": 4.9930069129672965e-05, "loss": 0.3184, "step": 422 }, { "epoch": 0.2966859547606523, "grad_norm": 19.460153579711914, "learning_rate": 4.992898124341296e-05, "loss": 0.2748, "step": 423 }, { "epoch": 0.2973873399964931, "grad_norm": 2.9695911407470703, "learning_rate": 4.9927884972471015e-05, "loss": 0.1925, "step": 424 }, { "epoch": 0.2980887252323339, "grad_norm": 12.33647346496582, "learning_rate": 4.9926780317215854e-05, "loss": 0.1908, "step": 425 }, { "epoch": 0.29879011046817466, "grad_norm": 2.8421826362609863, "learning_rate": 4.992566727801902e-05, "loss": 0.101, "step": 426 }, { "epoch": 0.29949149570401545, "grad_norm": 171.8658447265625, "learning_rate": 4.9924545855254875e-05, "loss": 3.9311, "step": 427 }, { "epoch": 0.30019288093985624, "grad_norm": 83.66468048095703, "learning_rate": 4.992341604930061e-05, "loss": 1.6806, "step": 428 }, { "epoch": 0.300894266175697, "grad_norm": 0.7609736323356628, "learning_rate": 4.9922277860536216e-05, "loss": 0.0041, "step": 429 }, { "epoch": 0.30159565141153777, "grad_norm": 4.077641487121582, "learning_rate": 4.9921131289344514e-05, "loss": 0.0418, "step": 430 }, { "epoch": 0.30229703664737856, "grad_norm": 3.2341866493225098, "learning_rate": 4.9919976336111155e-05, "loss": 0.0293, "step": 431 }, { "epoch": 0.30299842188321935, "grad_norm": 148.74095153808594, "learning_rate": 4.991881300122458e-05, "loss": 3.3951, "step": 432 }, { "epoch": 0.30369980711906014, "grad_norm": 77.130126953125, "learning_rate": 4.99176412850761e-05, "loss": 2.9407, "step": 433 }, { "epoch": 0.30440119235490093, "grad_norm": 162.98777770996094, "learning_rate": 4.991646118805978e-05, "loss": 4.3504, "step": 434 }, { "epoch": 0.3051025775907417, "grad_norm": 84.4205551147461, "learning_rate": 4.991527271057256e-05, "loss": 1.7142, "step": 435 }, { "epoch": 0.3058039628265825, "grad_norm": 0.8636782765388489, "learning_rate": 4.991407585301416e-05, "loss": 0.0053, "step": 436 }, { "epoch": 0.3065053480624233, "grad_norm": 1.0418014526367188, "learning_rate": 4.9912870615787144e-05, "loss": 0.0059, "step": 437 }, { "epoch": 0.3072067332982641, "grad_norm": 0.7050487399101257, "learning_rate": 4.991165699929688e-05, "loss": 0.0041, "step": 438 }, { "epoch": 0.3079081185341049, "grad_norm": 0.3087022304534912, "learning_rate": 4.991043500395157e-05, "loss": 0.0017, "step": 439 }, { "epoch": 0.30860950376994567, "grad_norm": 0.09886661171913147, "learning_rate": 4.990920463016221e-05, "loss": 0.0005, "step": 440 }, { "epoch": 0.3093108890057864, "grad_norm": 0.03563763573765755, "learning_rate": 4.9907965878342634e-05, "loss": 0.0002, "step": 441 }, { "epoch": 0.3100122742416272, "grad_norm": 79.55757141113281, "learning_rate": 4.990671874890948e-05, "loss": 2.966, "step": 442 }, { "epoch": 0.310713659477468, "grad_norm": 0.6562897562980652, "learning_rate": 4.990546324228222e-05, "loss": 0.0022, "step": 443 }, { "epoch": 0.3114150447133088, "grad_norm": 74.84564208984375, "learning_rate": 4.990419935888313e-05, "loss": 3.0154, "step": 444 }, { "epoch": 0.31211642994914957, "grad_norm": 0.03038967400789261, "learning_rate": 4.9902927099137305e-05, "loss": 0.0002, "step": 445 }, { "epoch": 0.31281781518499036, "grad_norm": 0.023526914417743683, "learning_rate": 4.9901646463472674e-05, "loss": 0.0001, "step": 446 }, { "epoch": 0.31351920042083115, "grad_norm": 0.08529358357191086, "learning_rate": 4.990035745231995e-05, "loss": 0.0005, "step": 447 }, { "epoch": 0.31422058565667194, "grad_norm": 0.11940905451774597, "learning_rate": 4.989906006611269e-05, "loss": 0.0007, "step": 448 }, { "epoch": 0.3149219708925127, "grad_norm": 99.68978118896484, "learning_rate": 4.989775430528727e-05, "loss": 1.6958, "step": 449 }, { "epoch": 0.3156233561283535, "grad_norm": 0.3344658315181732, "learning_rate": 4.9896440170282864e-05, "loss": 0.002, "step": 450 }, { "epoch": 0.3163247413641943, "grad_norm": 0.7184603810310364, "learning_rate": 4.9895117661541476e-05, "loss": 0.0043, "step": 451 }, { "epoch": 0.3170261266000351, "grad_norm": 0.5113491415977478, "learning_rate": 4.989378677950791e-05, "loss": 0.0031, "step": 452 }, { "epoch": 0.31772751183587583, "grad_norm": 0.49223193526268005, "learning_rate": 4.989244752462982e-05, "loss": 0.0029, "step": 453 }, { "epoch": 0.3184288970717166, "grad_norm": 64.00535583496094, "learning_rate": 4.989109989735763e-05, "loss": 1.2629, "step": 454 }, { "epoch": 0.3191302823075574, "grad_norm": 0.18930155038833618, "learning_rate": 4.988974389814463e-05, "loss": 0.0011, "step": 455 }, { "epoch": 0.3198316675433982, "grad_norm": 0.3088482916355133, "learning_rate": 4.9888379527446885e-05, "loss": 0.0018, "step": 456 }, { "epoch": 0.320533052779239, "grad_norm": 71.42636108398438, "learning_rate": 4.988700678572329e-05, "loss": 1.9691, "step": 457 }, { "epoch": 0.3212344380150798, "grad_norm": 0.20487411320209503, "learning_rate": 4.988562567343557e-05, "loss": 0.0012, "step": 458 }, { "epoch": 0.32193582325092057, "grad_norm": 73.37615203857422, "learning_rate": 4.988423619104823e-05, "loss": 1.8636, "step": 459 }, { "epoch": 0.32263720848676136, "grad_norm": 66.4114761352539, "learning_rate": 4.988283833902863e-05, "loss": 1.3073, "step": 460 }, { "epoch": 0.32333859372260215, "grad_norm": 4.2518744468688965, "learning_rate": 4.988143211784693e-05, "loss": 0.0318, "step": 461 }, { "epoch": 0.32403997895844294, "grad_norm": 4.782449245452881, "learning_rate": 4.9880017527976083e-05, "loss": 0.0384, "step": 462 }, { "epoch": 0.32474136419428373, "grad_norm": 22.517131805419922, "learning_rate": 4.9878594569891896e-05, "loss": 0.2973, "step": 463 }, { "epoch": 0.3254427494301245, "grad_norm": 5.63627290725708, "learning_rate": 4.987716324407296e-05, "loss": 0.0496, "step": 464 }, { "epoch": 0.32614413466596526, "grad_norm": 6.364257335662842, "learning_rate": 4.9875723551000696e-05, "loss": 0.0578, "step": 465 }, { "epoch": 0.32684551990180605, "grad_norm": 147.82330322265625, "learning_rate": 4.987427549115933e-05, "loss": 1.8188, "step": 466 }, { "epoch": 0.32754690513764684, "grad_norm": 3.7355575561523438, "learning_rate": 4.987281906503591e-05, "loss": 0.0274, "step": 467 }, { "epoch": 0.32824829037348763, "grad_norm": 0.6107962131500244, "learning_rate": 4.987135427312029e-05, "loss": 0.0038, "step": 468 }, { "epoch": 0.3289496756093284, "grad_norm": 61.26199722290039, "learning_rate": 4.9869881115905146e-05, "loss": 0.8915, "step": 469 }, { "epoch": 0.3296510608451692, "grad_norm": 0.3030995726585388, "learning_rate": 4.986839959388596e-05, "loss": 0.0018, "step": 470 }, { "epoch": 0.33035244608101, "grad_norm": 0.12271957844495773, "learning_rate": 4.986690970756104e-05, "loss": 0.0007, "step": 471 }, { "epoch": 0.3310538313168508, "grad_norm": 71.7486801147461, "learning_rate": 4.9865411457431485e-05, "loss": 1.8673, "step": 472 }, { "epoch": 0.3317552165526916, "grad_norm": 69.17851257324219, "learning_rate": 4.986390484400124e-05, "loss": 1.6587, "step": 473 }, { "epoch": 0.33245660178853237, "grad_norm": 0.18208004534244537, "learning_rate": 4.9862389867777023e-05, "loss": 0.001, "step": 474 }, { "epoch": 0.33315798702437316, "grad_norm": 122.65370178222656, "learning_rate": 4.986086652926839e-05, "loss": 0.9068, "step": 475 }, { "epoch": 0.33385937226021395, "grad_norm": 0.6119470596313477, "learning_rate": 4.9859334828987715e-05, "loss": 0.0039, "step": 476 }, { "epoch": 0.3345607574960547, "grad_norm": 4.483019828796387, "learning_rate": 4.985779476745017e-05, "loss": 0.0343, "step": 477 }, { "epoch": 0.3352621427318955, "grad_norm": 0.7592105269432068, "learning_rate": 4.985624634517374e-05, "loss": 0.005, "step": 478 }, { "epoch": 0.33596352796773626, "grad_norm": 0.6566614508628845, "learning_rate": 4.985468956267922e-05, "loss": 0.0037, "step": 479 }, { "epoch": 0.33666491320357705, "grad_norm": 67.39823150634766, "learning_rate": 4.985312442049023e-05, "loss": 1.7049, "step": 480 }, { "epoch": 0.33736629843941784, "grad_norm": 0.13803933560848236, "learning_rate": 4.98515509191332e-05, "loss": 0.0008, "step": 481 }, { "epoch": 0.33806768367525863, "grad_norm": 72.15523529052734, "learning_rate": 4.984996905913736e-05, "loss": 1.6634, "step": 482 }, { "epoch": 0.3387690689110994, "grad_norm": 0.10033612698316574, "learning_rate": 4.984837884103475e-05, "loss": 0.0005, "step": 483 }, { "epoch": 0.3394704541469402, "grad_norm": 62.519317626953125, "learning_rate": 4.984678026536024e-05, "loss": 1.2577, "step": 484 }, { "epoch": 0.340171839382781, "grad_norm": 0.36229056119918823, "learning_rate": 4.984517333265148e-05, "loss": 0.0023, "step": 485 }, { "epoch": 0.3408732246186218, "grad_norm": 0.8776547312736511, "learning_rate": 4.9843558043448985e-05, "loss": 0.0056, "step": 486 }, { "epoch": 0.3415746098544626, "grad_norm": 108.73171997070312, "learning_rate": 4.984193439829602e-05, "loss": 2.7267, "step": 487 }, { "epoch": 0.3422759950903034, "grad_norm": 0.6344736814498901, "learning_rate": 4.984030239773868e-05, "loss": 0.0041, "step": 488 }, { "epoch": 0.3429773803261441, "grad_norm": 113.62140655517578, "learning_rate": 4.98386620423259e-05, "loss": 1.7271, "step": 489 }, { "epoch": 0.3436787655619849, "grad_norm": 1.618384599685669, "learning_rate": 4.983701333260938e-05, "loss": 0.011, "step": 490 }, { "epoch": 0.3443801507978257, "grad_norm": 138.78854370117188, "learning_rate": 4.9835356269143676e-05, "loss": 1.9143, "step": 491 }, { "epoch": 0.3450815360336665, "grad_norm": 49.62142562866211, "learning_rate": 4.98336908524861e-05, "loss": 0.753, "step": 492 }, { "epoch": 0.34578292126950727, "grad_norm": 3.6228299140930176, "learning_rate": 4.9832017083196816e-05, "loss": 0.0277, "step": 493 }, { "epoch": 0.34648430650534806, "grad_norm": 24.915790557861328, "learning_rate": 4.98303349618388e-05, "loss": 0.356, "step": 494 }, { "epoch": 0.34718569174118885, "grad_norm": 7.880053997039795, "learning_rate": 4.9828644488977795e-05, "loss": 0.0727, "step": 495 }, { "epoch": 0.34788707697702964, "grad_norm": 12.95199203491211, "learning_rate": 4.982694566518239e-05, "loss": 0.25, "step": 496 }, { "epoch": 0.34858846221287043, "grad_norm": 2.059384822845459, "learning_rate": 4.982523849102397e-05, "loss": 0.2069, "step": 497 }, { "epoch": 0.3492898474487112, "grad_norm": 17.132720947265625, "learning_rate": 4.982352296707675e-05, "loss": 0.2093, "step": 498 }, { "epoch": 0.349991232684552, "grad_norm": 12.591998100280762, "learning_rate": 4.98217990939177e-05, "loss": 0.1889, "step": 499 }, { "epoch": 0.3506926179203928, "grad_norm": 12.747396469116211, "learning_rate": 4.982006687212666e-05, "loss": 0.1755, "step": 500 }, { "epoch": 0.35139400315623354, "grad_norm": 77.32592010498047, "learning_rate": 4.981832630228624e-05, "loss": 0.9517, "step": 501 }, { "epoch": 0.3520953883920743, "grad_norm": 8.452153205871582, "learning_rate": 4.9816577384981866e-05, "loss": 0.0913, "step": 502 }, { "epoch": 0.3527967736279151, "grad_norm": 2.1267127990722656, "learning_rate": 4.9814820120801776e-05, "loss": 0.0178, "step": 503 }, { "epoch": 0.3534981588637559, "grad_norm": 0.8161488771438599, "learning_rate": 4.981305451033701e-05, "loss": 0.0058, "step": 504 }, { "epoch": 0.3541995440995967, "grad_norm": 0.10489612072706223, "learning_rate": 4.981128055418143e-05, "loss": 0.0006, "step": 505 }, { "epoch": 0.3549009293354375, "grad_norm": 69.51275634765625, "learning_rate": 4.9809498252931674e-05, "loss": 2.62, "step": 506 }, { "epoch": 0.3556023145712783, "grad_norm": 0.004451240878552198, "learning_rate": 4.980770760718721e-05, "loss": 0.0, "step": 507 }, { "epoch": 0.35630369980711907, "grad_norm": 67.6517105102539, "learning_rate": 4.9805908617550335e-05, "loss": 3.25, "step": 508 }, { "epoch": 0.35700508504295986, "grad_norm": 0.016789492219686508, "learning_rate": 4.980410128462609e-05, "loss": 0.0001, "step": 509 }, { "epoch": 0.35770647027880065, "grad_norm": 0.025487884879112244, "learning_rate": 4.980228560902239e-05, "loss": 0.0001, "step": 510 }, { "epoch": 0.35840785551464144, "grad_norm": 0.042408283799886703, "learning_rate": 4.98004615913499e-05, "loss": 0.0002, "step": 511 }, { "epoch": 0.35910924075048223, "grad_norm": 0.038653161376714706, "learning_rate": 4.979862923222212e-05, "loss": 0.0002, "step": 512 }, { "epoch": 0.35981062598632296, "grad_norm": 0.03510050103068352, "learning_rate": 4.9796788532255356e-05, "loss": 0.0002, "step": 513 }, { "epoch": 0.36051201122216375, "grad_norm": 0.008441315963864326, "learning_rate": 4.979493949206872e-05, "loss": 0.0, "step": 514 }, { "epoch": 0.36121339645800454, "grad_norm": 0.006373463664203882, "learning_rate": 4.979308211228411e-05, "loss": 0.0, "step": 515 }, { "epoch": 0.36191478169384533, "grad_norm": 79.59523010253906, "learning_rate": 4.979121639352625e-05, "loss": 2.9487, "step": 516 }, { "epoch": 0.3626161669296861, "grad_norm": 153.0602264404297, "learning_rate": 4.9789342336422665e-05, "loss": 3.6533, "step": 517 }, { "epoch": 0.3633175521655269, "grad_norm": 0.7085576057434082, "learning_rate": 4.9787459941603666e-05, "loss": 0.0042, "step": 518 }, { "epoch": 0.3640189374013677, "grad_norm": 155.89154052734375, "learning_rate": 4.9785569209702395e-05, "loss": 7.9018, "step": 519 }, { "epoch": 0.3647203226372085, "grad_norm": 5.253132343292236, "learning_rate": 4.9783670141354776e-05, "loss": 0.042, "step": 520 }, { "epoch": 0.3654217078730493, "grad_norm": 26.310007095336914, "learning_rate": 4.978176273719957e-05, "loss": 0.2762, "step": 521 }, { "epoch": 0.3661230931088901, "grad_norm": 9.570679664611816, "learning_rate": 4.977984699787829e-05, "loss": 0.0898, "step": 522 }, { "epoch": 0.36682447834473086, "grad_norm": 214.85060119628906, "learning_rate": 4.977792292403529e-05, "loss": 6.7239, "step": 523 }, { "epoch": 0.36752586358057165, "grad_norm": 2.7513771057128906, "learning_rate": 4.977599051631773e-05, "loss": 0.0228, "step": 524 }, { "epoch": 0.3682272488164124, "grad_norm": 33.629356384277344, "learning_rate": 4.977404977537554e-05, "loss": 0.4374, "step": 525 }, { "epoch": 0.3689286340522532, "grad_norm": 3.8864808082580566, "learning_rate": 4.97721007018615e-05, "loss": 0.0308, "step": 526 }, { "epoch": 0.36963001928809397, "grad_norm": 1.0654228925704956, "learning_rate": 4.9770143296431144e-05, "loss": 0.0077, "step": 527 }, { "epoch": 0.37033140452393476, "grad_norm": 131.74990844726562, "learning_rate": 4.9768177559742844e-05, "loss": 4.0998, "step": 528 }, { "epoch": 0.37103278975977555, "grad_norm": 0.10420957952737808, "learning_rate": 4.976620349245776e-05, "loss": 0.0006, "step": 529 }, { "epoch": 0.37173417499561634, "grad_norm": 72.78388977050781, "learning_rate": 4.976422109523985e-05, "loss": 3.8242, "step": 530 }, { "epoch": 0.37243556023145713, "grad_norm": 69.0342025756836, "learning_rate": 4.976223036875588e-05, "loss": 1.9872, "step": 531 }, { "epoch": 0.3731369454672979, "grad_norm": 0.43323075771331787, "learning_rate": 4.9760231313675426e-05, "loss": 0.0028, "step": 532 }, { "epoch": 0.3738383307031387, "grad_norm": 1.0747731924057007, "learning_rate": 4.975822393067084e-05, "loss": 0.0061, "step": 533 }, { "epoch": 0.3745397159389795, "grad_norm": 69.71436309814453, "learning_rate": 4.9756208220417304e-05, "loss": 1.8796, "step": 534 }, { "epoch": 0.3752411011748203, "grad_norm": 127.05543518066406, "learning_rate": 4.975418418359278e-05, "loss": 2.8441, "step": 535 }, { "epoch": 0.3759424864106611, "grad_norm": 0.9959923624992371, "learning_rate": 4.975215182087804e-05, "loss": 0.0065, "step": 536 }, { "epoch": 0.3766438716465018, "grad_norm": 2.1746952533721924, "learning_rate": 4.9750111132956654e-05, "loss": 0.0156, "step": 537 }, { "epoch": 0.3773452568823426, "grad_norm": 92.57381439208984, "learning_rate": 4.9748062120514996e-05, "loss": 1.4085, "step": 538 }, { "epoch": 0.3780466421181834, "grad_norm": 75.45511627197266, "learning_rate": 4.974600478424223e-05, "loss": 0.9909, "step": 539 }, { "epoch": 0.3787480273540242, "grad_norm": 25.877460479736328, "learning_rate": 4.974393912483034e-05, "loss": 0.3578, "step": 540 }, { "epoch": 0.379449412589865, "grad_norm": 12.262901306152344, "learning_rate": 4.9741865142974076e-05, "loss": 0.1338, "step": 541 }, { "epoch": 0.38015079782570577, "grad_norm": 10.54548168182373, "learning_rate": 4.973978283937102e-05, "loss": 0.2111, "step": 542 }, { "epoch": 0.38085218306154656, "grad_norm": 35.909976959228516, "learning_rate": 4.973769221472153e-05, "loss": 0.5549, "step": 543 }, { "epoch": 0.38155356829738735, "grad_norm": 13.012696266174316, "learning_rate": 4.973559326972878e-05, "loss": 0.1784, "step": 544 }, { "epoch": 0.38225495353322814, "grad_norm": 16.022974014282227, "learning_rate": 4.973348600509874e-05, "loss": 0.2235, "step": 545 }, { "epoch": 0.3829563387690689, "grad_norm": 22.86570167541504, "learning_rate": 4.973137042154016e-05, "loss": 0.3561, "step": 546 }, { "epoch": 0.3836577240049097, "grad_norm": 13.543822288513184, "learning_rate": 4.972924651976461e-05, "loss": 0.3077, "step": 547 }, { "epoch": 0.3843591092407505, "grad_norm": 17.42881965637207, "learning_rate": 4.972711430048644e-05, "loss": 0.2779, "step": 548 }, { "epoch": 0.38506049447659124, "grad_norm": 19.7313175201416, "learning_rate": 4.972497376442282e-05, "loss": 0.2582, "step": 549 }, { "epoch": 0.38576187971243203, "grad_norm": 12.354894638061523, "learning_rate": 4.9722824912293695e-05, "loss": 0.2374, "step": 550 }, { "epoch": 0.3864632649482728, "grad_norm": 17.16146469116211, "learning_rate": 4.972066774482182e-05, "loss": 0.2085, "step": 551 }, { "epoch": 0.3871646501841136, "grad_norm": 49.288734436035156, "learning_rate": 4.971850226273274e-05, "loss": 0.6211, "step": 552 }, { "epoch": 0.3878660354199544, "grad_norm": 4.870689392089844, "learning_rate": 4.97163284667548e-05, "loss": 0.0413, "step": 553 }, { "epoch": 0.3885674206557952, "grad_norm": 4.268405437469482, "learning_rate": 4.971414635761914e-05, "loss": 0.0354, "step": 554 }, { "epoch": 0.389268805891636, "grad_norm": 2.800025701522827, "learning_rate": 4.971195593605969e-05, "loss": 0.0215, "step": 555 }, { "epoch": 0.3899701911274768, "grad_norm": 136.0956573486328, "learning_rate": 4.9709757202813195e-05, "loss": 3.3931, "step": 556 }, { "epoch": 0.39067157636331756, "grad_norm": 68.56623840332031, "learning_rate": 4.970755015861918e-05, "loss": 1.8678, "step": 557 }, { "epoch": 0.39137296159915835, "grad_norm": 68.25215911865234, "learning_rate": 4.970533480421996e-05, "loss": 1.9984, "step": 558 }, { "epoch": 0.39207434683499914, "grad_norm": 69.02233123779297, "learning_rate": 4.9703111140360656e-05, "loss": 2.0662, "step": 559 }, { "epoch": 0.39277573207083993, "grad_norm": 61.60824203491211, "learning_rate": 4.970087916778918e-05, "loss": 1.5281, "step": 560 }, { "epoch": 0.39347711730668067, "grad_norm": 48.50178146362305, "learning_rate": 4.9698638887256254e-05, "loss": 0.8115, "step": 561 }, { "epoch": 0.39417850254252146, "grad_norm": 41.32649612426758, "learning_rate": 4.969639029951537e-05, "loss": 0.6374, "step": 562 }, { "epoch": 0.39487988777836225, "grad_norm": 6.946910858154297, "learning_rate": 4.9694133405322815e-05, "loss": 0.2402, "step": 563 }, { "epoch": 0.39558127301420304, "grad_norm": 22.8358097076416, "learning_rate": 4.9691868205437694e-05, "loss": 0.3883, "step": 564 }, { "epoch": 0.39628265825004383, "grad_norm": 15.402759552001953, "learning_rate": 4.968959470062189e-05, "loss": 0.7708, "step": 565 }, { "epoch": 0.3969840434858846, "grad_norm": 28.08781623840332, "learning_rate": 4.968731289164006e-05, "loss": 0.7703, "step": 566 }, { "epoch": 0.3976854287217254, "grad_norm": 26.910886764526367, "learning_rate": 4.9685022779259696e-05, "loss": 0.6299, "step": 567 }, { "epoch": 0.3983868139575662, "grad_norm": 24.341957092285156, "learning_rate": 4.968272436425105e-05, "loss": 0.5751, "step": 568 }, { "epoch": 0.399088199193407, "grad_norm": 23.469263076782227, "learning_rate": 4.9680417647387186e-05, "loss": 0.4664, "step": 569 }, { "epoch": 0.3997895844292478, "grad_norm": 15.347972869873047, "learning_rate": 4.967810262944394e-05, "loss": 0.2252, "step": 570 }, { "epoch": 0.40049096966508857, "grad_norm": 5.6460280418396, "learning_rate": 4.967577931119996e-05, "loss": 0.2232, "step": 571 }, { "epoch": 0.40119235490092936, "grad_norm": 4.2383856773376465, "learning_rate": 4.967344769343667e-05, "loss": 0.0396, "step": 572 }, { "epoch": 0.4018937401367701, "grad_norm": 1.6932095289230347, "learning_rate": 4.96711077769383e-05, "loss": 0.0137, "step": 573 }, { "epoch": 0.4025951253726109, "grad_norm": 0.41152113676071167, "learning_rate": 4.966875956249186e-05, "loss": 0.0029, "step": 574 }, { "epoch": 0.4032965106084517, "grad_norm": 0.06895897537469864, "learning_rate": 4.9666403050887155e-05, "loss": 0.0004, "step": 575 }, { "epoch": 0.40399789584429247, "grad_norm": 0.004564318805932999, "learning_rate": 4.9664038242916786e-05, "loss": 0.0, "step": 576 }, { "epoch": 0.40469928108013326, "grad_norm": 0.0019707169849425554, "learning_rate": 4.966166513937613e-05, "loss": 0.0, "step": 577 }, { "epoch": 0.40540066631597405, "grad_norm": 0.00037522497586905956, "learning_rate": 4.965928374106337e-05, "loss": 0.0, "step": 578 }, { "epoch": 0.40610205155181484, "grad_norm": 5.6252967624459416e-05, "learning_rate": 4.965689404877946e-05, "loss": 0.0, "step": 579 }, { "epoch": 0.4068034367876556, "grad_norm": 1.1128309779451229e-05, "learning_rate": 4.9654496063328174e-05, "loss": 0.0, "step": 580 }, { "epoch": 0.4075048220234964, "grad_norm": 119.89021301269531, "learning_rate": 4.965208978551605e-05, "loss": 10.8344, "step": 581 }, { "epoch": 0.4082062072593372, "grad_norm": 1.3851165931555443e-05, "learning_rate": 4.964967521615241e-05, "loss": 0.0, "step": 582 }, { "epoch": 0.408907592495178, "grad_norm": 3.7249556044116616e-06, "learning_rate": 4.96472523560494e-05, "loss": 0.0, "step": 583 }, { "epoch": 0.4096089777310188, "grad_norm": 2.917176971095614e-06, "learning_rate": 4.9644821206021906e-05, "loss": 0.0, "step": 584 }, { "epoch": 0.4103103629668595, "grad_norm": 3.3871406230900902e-06, "learning_rate": 4.964238176688765e-05, "loss": 0.0, "step": 585 }, { "epoch": 0.4110117482027003, "grad_norm": 2.260056135128252e-06, "learning_rate": 4.9639934039467106e-05, "loss": 0.0, "step": 586 }, { "epoch": 0.4117131334385411, "grad_norm": 1.5038928040667088e-06, "learning_rate": 4.963747802458355e-05, "loss": 0.0, "step": 587 }, { "epoch": 0.4124145186743819, "grad_norm": 59.293521881103516, "learning_rate": 4.963501372306306e-05, "loss": 5.3994, "step": 588 }, { "epoch": 0.4131159039102227, "grad_norm": 59.763816833496094, "learning_rate": 4.963254113573446e-05, "loss": 5.6505, "step": 589 }, { "epoch": 0.41381728914606347, "grad_norm": 132.7192840576172, "learning_rate": 4.963006026342941e-05, "loss": 9.1499, "step": 590 }, { "epoch": 0.41451867438190426, "grad_norm": 0.00024285895051434636, "learning_rate": 4.962757110698232e-05, "loss": 0.0, "step": 591 }, { "epoch": 0.41522005961774505, "grad_norm": 0.0009618580807000399, "learning_rate": 4.9625073667230404e-05, "loss": 0.0, "step": 592 }, { "epoch": 0.41592144485358584, "grad_norm": 0.004245178773999214, "learning_rate": 4.962256794501365e-05, "loss": 0.0, "step": 593 }, { "epoch": 0.41662283008942663, "grad_norm": 0.01505737192928791, "learning_rate": 4.962005394117486e-05, "loss": 0.0001, "step": 594 }, { "epoch": 0.4173242153252674, "grad_norm": 0.01870640181005001, "learning_rate": 4.961753165655958e-05, "loss": 0.0001, "step": 595 }, { "epoch": 0.4180256005611082, "grad_norm": 0.047609515488147736, "learning_rate": 4.961500109201618e-05, "loss": 0.0003, "step": 596 }, { "epoch": 0.41872698579694895, "grad_norm": 0.15332278609275818, "learning_rate": 4.9612462248395775e-05, "loss": 0.001, "step": 597 }, { "epoch": 0.41942837103278974, "grad_norm": 113.53760528564453, "learning_rate": 4.96099151265523e-05, "loss": 2.3303, "step": 598 }, { "epoch": 0.4201297562686305, "grad_norm": 0.6282661557197571, "learning_rate": 4.9607359727342465e-05, "loss": 0.0044, "step": 599 }, { "epoch": 0.4208311415044713, "grad_norm": 43.630615234375, "learning_rate": 4.9604796051625756e-05, "loss": 0.6733, "step": 600 }, { "epoch": 0.4215325267403121, "grad_norm": 72.0724868774414, "learning_rate": 4.960222410026444e-05, "loss": 1.4145, "step": 601 }, { "epoch": 0.4222339119761529, "grad_norm": 6.166507244110107, "learning_rate": 4.959964387412358e-05, "loss": 0.0608, "step": 602 }, { "epoch": 0.4229352972119937, "grad_norm": 80.03913879394531, "learning_rate": 4.9597055374071024e-05, "loss": 1.2555, "step": 603 }, { "epoch": 0.4236366824478345, "grad_norm": 17.456140518188477, "learning_rate": 4.959445860097738e-05, "loss": 0.238, "step": 604 }, { "epoch": 0.42433806768367527, "grad_norm": 36.63374328613281, "learning_rate": 4.9591853555716064e-05, "loss": 0.8324, "step": 605 }, { "epoch": 0.42503945291951606, "grad_norm": 16.27304458618164, "learning_rate": 4.9589240239163255e-05, "loss": 0.2125, "step": 606 }, { "epoch": 0.42574083815535685, "grad_norm": 19.427776336669922, "learning_rate": 4.9586618652197934e-05, "loss": 0.2701, "step": 607 }, { "epoch": 0.42644222339119764, "grad_norm": 8.516806602478027, "learning_rate": 4.958398879570185e-05, "loss": 0.3408, "step": 608 }, { "epoch": 0.4271436086270384, "grad_norm": 1.857074499130249, "learning_rate": 4.9581350670559524e-05, "loss": 0.2039, "step": 609 }, { "epoch": 0.42784499386287916, "grad_norm": 8.925201416015625, "learning_rate": 4.9578704277658286e-05, "loss": 0.09, "step": 610 }, { "epoch": 0.42854637909871995, "grad_norm": 4.490631580352783, "learning_rate": 4.957604961788823e-05, "loss": 0.0384, "step": 611 }, { "epoch": 0.42924776433456074, "grad_norm": 0.866288959980011, "learning_rate": 4.957338669214222e-05, "loss": 0.0063, "step": 612 }, { "epoch": 0.42994914957040153, "grad_norm": 0.6116040349006653, "learning_rate": 4.9570715501315925e-05, "loss": 0.0035, "step": 613 }, { "epoch": 0.4306505348062423, "grad_norm": 0.014350208453834057, "learning_rate": 4.956803604630777e-05, "loss": 0.0001, "step": 614 }, { "epoch": 0.4313519200420831, "grad_norm": 0.004978867247700691, "learning_rate": 4.956534832801897e-05, "loss": 0.0, "step": 615 }, { "epoch": 0.4320533052779239, "grad_norm": 0.0008859955123625696, "learning_rate": 4.9562652347353535e-05, "loss": 0.0, "step": 616 }, { "epoch": 0.4327546905137647, "grad_norm": 0.0007402606424875557, "learning_rate": 4.955994810521822e-05, "loss": 0.0, "step": 617 }, { "epoch": 0.4334560757496055, "grad_norm": 60.98609161376953, "learning_rate": 4.955723560252259e-05, "loss": 4.6234, "step": 618 }, { "epoch": 0.4341574609854463, "grad_norm": 60.63393783569336, "learning_rate": 4.955451484017896e-05, "loss": 4.669, "step": 619 }, { "epoch": 0.43485884622128707, "grad_norm": 0.00016736971156205982, "learning_rate": 4.955178581910246e-05, "loss": 0.0, "step": 620 }, { "epoch": 0.4355602314571278, "grad_norm": 0.00026081904070451856, "learning_rate": 4.9549048540210963e-05, "loss": 0.0, "step": 621 }, { "epoch": 0.4362616166929686, "grad_norm": 0.0002771130239125341, "learning_rate": 4.954630300442513e-05, "loss": 0.0, "step": 622 }, { "epoch": 0.4369630019288094, "grad_norm": 0.00024890206987038255, "learning_rate": 4.954354921266841e-05, "loss": 0.0, "step": 623 }, { "epoch": 0.43766438716465017, "grad_norm": 0.0003540183824952692, "learning_rate": 4.954078716586702e-05, "loss": 0.0, "step": 624 }, { "epoch": 0.43836577240049096, "grad_norm": 61.1258659362793, "learning_rate": 4.9538016864949953e-05, "loss": 4.1063, "step": 625 }, { "epoch": 0.43906715763633175, "grad_norm": 61.82120895385742, "learning_rate": 4.953523831084897e-05, "loss": 3.2575, "step": 626 }, { "epoch": 0.43976854287217254, "grad_norm": 0.009341489523649216, "learning_rate": 4.9532451504498644e-05, "loss": 0.0001, "step": 627 }, { "epoch": 0.44046992810801333, "grad_norm": 126.25260925292969, "learning_rate": 4.952965644683627e-05, "loss": 4.8894, "step": 628 }, { "epoch": 0.4411713133438541, "grad_norm": 64.0586166381836, "learning_rate": 4.9526853138801954e-05, "loss": 2.0939, "step": 629 }, { "epoch": 0.4418726985796949, "grad_norm": 66.46090698242188, "learning_rate": 4.952404158133857e-05, "loss": 2.1672, "step": 630 }, { "epoch": 0.4425740838155357, "grad_norm": 8.158260345458984, "learning_rate": 4.952122177539177e-05, "loss": 0.089, "step": 631 }, { "epoch": 0.4432754690513765, "grad_norm": 8.28303337097168, "learning_rate": 4.951839372190996e-05, "loss": 0.1792, "step": 632 }, { "epoch": 0.4439768542872172, "grad_norm": 6.670780658721924, "learning_rate": 4.951555742184435e-05, "loss": 0.3495, "step": 633 }, { "epoch": 0.444678239523058, "grad_norm": 18.7882022857666, "learning_rate": 4.95127128761489e-05, "loss": 0.2761, "step": 634 }, { "epoch": 0.4453796247588988, "grad_norm": 3.4713053703308105, "learning_rate": 4.950986008578037e-05, "loss": 0.2429, "step": 635 }, { "epoch": 0.4460810099947396, "grad_norm": 8.415278434753418, "learning_rate": 4.950699905169825e-05, "loss": 0.0883, "step": 636 }, { "epoch": 0.4467823952305804, "grad_norm": 5.512305736541748, "learning_rate": 4.950412977486484e-05, "loss": 0.0515, "step": 637 }, { "epoch": 0.4474837804664212, "grad_norm": 1.983485460281372, "learning_rate": 4.95012522562452e-05, "loss": 0.0162, "step": 638 }, { "epoch": 0.44818516570226197, "grad_norm": 0.5145743489265442, "learning_rate": 4.949836649680717e-05, "loss": 0.0038, "step": 639 }, { "epoch": 0.44888655093810276, "grad_norm": 0.0464518703520298, "learning_rate": 4.949547249752134e-05, "loss": 0.0003, "step": 640 }, { "epoch": 0.44958793617394355, "grad_norm": 63.47880935668945, "learning_rate": 4.9492570259361094e-05, "loss": 2.606, "step": 641 }, { "epoch": 0.45028932140978434, "grad_norm": 63.044219970703125, "learning_rate": 4.948965978330258e-05, "loss": 3.3714, "step": 642 }, { "epoch": 0.45099070664562513, "grad_norm": 61.190128326416016, "learning_rate": 4.948674107032471e-05, "loss": 3.3129, "step": 643 }, { "epoch": 0.4516920918814659, "grad_norm": 59.03745651245117, "learning_rate": 4.948381412140917e-05, "loss": 2.3587, "step": 644 }, { "epoch": 0.4523934771173067, "grad_norm": 61.125858306884766, "learning_rate": 4.948087893754043e-05, "loss": 2.7351, "step": 645 }, { "epoch": 0.45309486235314744, "grad_norm": 60.69954299926758, "learning_rate": 4.947793551970571e-05, "loss": 2.2391, "step": 646 }, { "epoch": 0.45379624758898823, "grad_norm": 56.91267395019531, "learning_rate": 4.9474983868895e-05, "loss": 1.5786, "step": 647 }, { "epoch": 0.454497632824829, "grad_norm": 0.8762818574905396, "learning_rate": 4.947202398610108e-05, "loss": 0.0067, "step": 648 }, { "epoch": 0.4551990180606698, "grad_norm": 2.4974522590637207, "learning_rate": 4.9469055872319483e-05, "loss": 0.0224, "step": 649 }, { "epoch": 0.4559004032965106, "grad_norm": 22.98019790649414, "learning_rate": 4.9466079528548505e-05, "loss": 0.3665, "step": 650 }, { "epoch": 0.4566017885323514, "grad_norm": 16.03438949584961, "learning_rate": 4.9463094955789214e-05, "loss": 0.315, "step": 651 }, { "epoch": 0.4573031737681922, "grad_norm": 8.860062599182129, "learning_rate": 4.946010215504546e-05, "loss": 0.1067, "step": 652 }, { "epoch": 0.458004559004033, "grad_norm": 5.521644115447998, "learning_rate": 4.945710112732386e-05, "loss": 0.2473, "step": 653 }, { "epoch": 0.45870594423987376, "grad_norm": 15.16481876373291, "learning_rate": 4.945409187363376e-05, "loss": 0.2257, "step": 654 }, { "epoch": 0.45940732947571455, "grad_norm": 15.714698791503906, "learning_rate": 4.945107439498732e-05, "loss": 0.2359, "step": 655 }, { "epoch": 0.46010871471155534, "grad_norm": 6.4135589599609375, "learning_rate": 4.944804869239944e-05, "loss": 0.2336, "step": 656 }, { "epoch": 0.46081009994739613, "grad_norm": 9.600846290588379, "learning_rate": 4.94450147668878e-05, "loss": 0.1174, "step": 657 }, { "epoch": 0.46151148518323687, "grad_norm": 7.477170467376709, "learning_rate": 4.944197261947283e-05, "loss": 0.0857, "step": 658 }, { "epoch": 0.46221287041907766, "grad_norm": 2.0777835845947266, "learning_rate": 4.943892225117774e-05, "loss": 0.0196, "step": 659 }, { "epoch": 0.46291425565491845, "grad_norm": 41.072837829589844, "learning_rate": 4.9435863663028495e-05, "loss": 0.7781, "step": 660 }, { "epoch": 0.46361564089075924, "grad_norm": 0.5752171277999878, "learning_rate": 4.943279685605383e-05, "loss": 0.0045, "step": 661 }, { "epoch": 0.46431702612660003, "grad_norm": 51.6488151550293, "learning_rate": 4.9429721831285246e-05, "loss": 1.1945, "step": 662 }, { "epoch": 0.4650184113624408, "grad_norm": 0.11998272687196732, "learning_rate": 4.942663858975701e-05, "loss": 0.0009, "step": 663 }, { "epoch": 0.4657197965982816, "grad_norm": 57.40029525756836, "learning_rate": 4.9423547132506135e-05, "loss": 1.7823, "step": 664 }, { "epoch": 0.4664211818341224, "grad_norm": 0.07954677194356918, "learning_rate": 4.942044746057242e-05, "loss": 0.0006, "step": 665 }, { "epoch": 0.4671225670699632, "grad_norm": 77.24893951416016, "learning_rate": 4.941733957499841e-05, "loss": 3.3749, "step": 666 }, { "epoch": 0.467823952305804, "grad_norm": 57.04423522949219, "learning_rate": 4.9414223476829434e-05, "loss": 1.6771, "step": 667 }, { "epoch": 0.46852533754164477, "grad_norm": 0.14917835593223572, "learning_rate": 4.941109916711355e-05, "loss": 0.0011, "step": 668 }, { "epoch": 0.46922672277748556, "grad_norm": 47.98805236816406, "learning_rate": 4.940796664690161e-05, "loss": 1.0673, "step": 669 }, { "epoch": 0.4699281080133263, "grad_norm": 63.10405349731445, "learning_rate": 4.940482591724721e-05, "loss": 1.7249, "step": 670 }, { "epoch": 0.4706294932491671, "grad_norm": 47.01327133178711, "learning_rate": 4.940167697920671e-05, "loss": 0.9642, "step": 671 }, { "epoch": 0.4713308784850079, "grad_norm": 4.804792881011963, "learning_rate": 4.9398519833839234e-05, "loss": 0.0514, "step": 672 }, { "epoch": 0.47203226372084867, "grad_norm": 46.21573257446289, "learning_rate": 4.939535448220667e-05, "loss": 0.7498, "step": 673 }, { "epoch": 0.47273364895668946, "grad_norm": 12.326519966125488, "learning_rate": 4.939218092537365e-05, "loss": 0.1734, "step": 674 }, { "epoch": 0.47343503419253025, "grad_norm": 2.402872085571289, "learning_rate": 4.938899916440759e-05, "loss": 0.2545, "step": 675 }, { "epoch": 0.47413641942837104, "grad_norm": 4.058804512023926, "learning_rate": 4.938580920037864e-05, "loss": 0.2573, "step": 676 }, { "epoch": 0.4748378046642118, "grad_norm": 11.151541709899902, "learning_rate": 4.938261103435973e-05, "loss": 0.4285, "step": 677 }, { "epoch": 0.4755391899000526, "grad_norm": 6.957059860229492, "learning_rate": 4.937940466742653e-05, "loss": 0.2347, "step": 678 }, { "epoch": 0.4762405751358934, "grad_norm": 5.085255146026611, "learning_rate": 4.9376190100657495e-05, "loss": 0.2671, "step": 679 }, { "epoch": 0.4769419603717342, "grad_norm": 9.267155647277832, "learning_rate": 4.93729673351338e-05, "loss": 0.1125, "step": 680 }, { "epoch": 0.477643345607575, "grad_norm": 6.162266731262207, "learning_rate": 4.9369736371939426e-05, "loss": 0.0683, "step": 681 }, { "epoch": 0.4783447308434157, "grad_norm": 3.6411261558532715, "learning_rate": 4.936649721216105e-05, "loss": 0.0356, "step": 682 }, { "epoch": 0.4790461160792565, "grad_norm": 48.720794677734375, "learning_rate": 4.9363249856888165e-05, "loss": 1.0375, "step": 683 }, { "epoch": 0.4797475013150973, "grad_norm": 0.3436586856842041, "learning_rate": 4.9359994307212994e-05, "loss": 0.0027, "step": 684 }, { "epoch": 0.4804488865509381, "grad_norm": 52.97607421875, "learning_rate": 4.93567305642305e-05, "loss": 1.6276, "step": 685 }, { "epoch": 0.4811502717867789, "grad_norm": 0.09510497748851776, "learning_rate": 4.935345862903844e-05, "loss": 0.0007, "step": 686 }, { "epoch": 0.4818516570226197, "grad_norm": 56.231292724609375, "learning_rate": 4.93501785027373e-05, "loss": 2.0342, "step": 687 }, { "epoch": 0.48255304225846046, "grad_norm": 0.21096964180469513, "learning_rate": 4.934689018643032e-05, "loss": 0.0015, "step": 688 }, { "epoch": 0.48325442749430125, "grad_norm": 0.19893679022789001, "learning_rate": 4.93435936812235e-05, "loss": 0.0014, "step": 689 }, { "epoch": 0.48395581273014204, "grad_norm": 51.99702072143555, "learning_rate": 4.9340288988225594e-05, "loss": 1.0089, "step": 690 }, { "epoch": 0.48465719796598283, "grad_norm": 53.519405364990234, "learning_rate": 4.9336976108548126e-05, "loss": 1.7908, "step": 691 }, { "epoch": 0.4853585832018236, "grad_norm": 106.84630584716797, "learning_rate": 4.9333655043305354e-05, "loss": 3.6845, "step": 692 }, { "epoch": 0.4860599684376644, "grad_norm": 0.24259038269519806, "learning_rate": 4.9330325793614285e-05, "loss": 0.0019, "step": 693 }, { "epoch": 0.48676135367350515, "grad_norm": 0.8419340252876282, "learning_rate": 4.93269883605947e-05, "loss": 0.0072, "step": 694 }, { "epoch": 0.48746273890934594, "grad_norm": 1.204994559288025, "learning_rate": 4.932364274536912e-05, "loss": 0.0087, "step": 695 }, { "epoch": 0.48816412414518673, "grad_norm": 1.1827445030212402, "learning_rate": 4.932028894906279e-05, "loss": 0.0104, "step": 696 }, { "epoch": 0.4888655093810275, "grad_norm": 0.5520339608192444, "learning_rate": 4.931692697280378e-05, "loss": 0.0045, "step": 697 }, { "epoch": 0.4895668946168683, "grad_norm": 0.9636574983596802, "learning_rate": 4.931355681772283e-05, "loss": 0.0079, "step": 698 }, { "epoch": 0.4902682798527091, "grad_norm": 0.5019407868385315, "learning_rate": 4.931017848495348e-05, "loss": 0.004, "step": 699 }, { "epoch": 0.4909696650885499, "grad_norm": 0.0757029727101326, "learning_rate": 4.9306791975632026e-05, "loss": 0.0006, "step": 700 }, { "epoch": 0.4916710503243907, "grad_norm": 110.13203430175781, "learning_rate": 4.9303397290897465e-05, "loss": 4.3731, "step": 701 }, { "epoch": 0.49237243556023147, "grad_norm": 0.03152499720454216, "learning_rate": 4.929999443189159e-05, "loss": 0.0002, "step": 702 }, { "epoch": 0.49307382079607226, "grad_norm": 0.029290443286299706, "learning_rate": 4.929658339975893e-05, "loss": 0.0002, "step": 703 }, { "epoch": 0.49377520603191305, "grad_norm": 0.015484875068068504, "learning_rate": 4.9293164195646754e-05, "loss": 0.0001, "step": 704 }, { "epoch": 0.49447659126775384, "grad_norm": 53.61540222167969, "learning_rate": 4.928973682070509e-05, "loss": 2.7716, "step": 705 }, { "epoch": 0.4951779765035946, "grad_norm": 55.3106575012207, "learning_rate": 4.9286301276086715e-05, "loss": 2.3966, "step": 706 }, { "epoch": 0.49587936173943536, "grad_norm": 0.014001145958900452, "learning_rate": 4.928285756294714e-05, "loss": 0.0001, "step": 707 }, { "epoch": 0.49658074697527615, "grad_norm": 0.026434065774083138, "learning_rate": 4.927940568244465e-05, "loss": 0.0002, "step": 708 }, { "epoch": 0.49728213221111695, "grad_norm": 0.04528193548321724, "learning_rate": 4.927594563574023e-05, "loss": 0.0003, "step": 709 }, { "epoch": 0.49798351744695774, "grad_norm": 0.0950818806886673, "learning_rate": 4.9272477423997666e-05, "loss": 0.0007, "step": 710 }, { "epoch": 0.4986849026827985, "grad_norm": 0.1095154732465744, "learning_rate": 4.926900104838346e-05, "loss": 0.0008, "step": 711 }, { "epoch": 0.4993862879186393, "grad_norm": 0.13107261061668396, "learning_rate": 4.9265516510066864e-05, "loss": 0.001, "step": 712 }, { "epoch": 0.5000876731544801, "grad_norm": 53.58122634887695, "learning_rate": 4.926202381021988e-05, "loss": 1.8605, "step": 713 }, { "epoch": 0.5007890583903208, "grad_norm": 48.86884689331055, "learning_rate": 4.925852295001725e-05, "loss": 1.3723, "step": 714 }, { "epoch": 0.5014904436261617, "grad_norm": 0.40202492475509644, "learning_rate": 4.9255013930636465e-05, "loss": 0.0033, "step": 715 }, { "epoch": 0.5021918288620024, "grad_norm": 43.420196533203125, "learning_rate": 4.925149675325775e-05, "loss": 0.9394, "step": 716 }, { "epoch": 0.5028932140978433, "grad_norm": 56.66259002685547, "learning_rate": 4.9247971419064097e-05, "loss": 1.2755, "step": 717 }, { "epoch": 0.503594599333684, "grad_norm": 3.452730894088745, "learning_rate": 4.9244437929241215e-05, "loss": 0.0353, "step": 718 }, { "epoch": 0.5042959845695248, "grad_norm": 4.454288005828857, "learning_rate": 4.924089628497757e-05, "loss": 0.0486, "step": 719 }, { "epoch": 0.5049973698053656, "grad_norm": 18.87127685546875, "learning_rate": 4.923734648746437e-05, "loss": 0.3667, "step": 720 }, { "epoch": 0.5056987550412064, "grad_norm": 11.85842514038086, "learning_rate": 4.923378853789556e-05, "loss": 0.3075, "step": 721 }, { "epoch": 0.5064001402770472, "grad_norm": 9.63955307006836, "learning_rate": 4.923022243746783e-05, "loss": 0.1304, "step": 722 }, { "epoch": 0.507101525512888, "grad_norm": 9.134459495544434, "learning_rate": 4.922664818738063e-05, "loss": 0.1346, "step": 723 }, { "epoch": 0.5078029107487287, "grad_norm": 2.0840797424316406, "learning_rate": 4.92230657888361e-05, "loss": 0.2203, "step": 724 }, { "epoch": 0.5085042959845695, "grad_norm": 8.857693672180176, "learning_rate": 4.921947524303919e-05, "loss": 0.1232, "step": 725 }, { "epoch": 0.5092056812204103, "grad_norm": 8.82702350616455, "learning_rate": 4.921587655119753e-05, "loss": 0.1143, "step": 726 }, { "epoch": 0.5099070664562511, "grad_norm": 4.017044544219971, "learning_rate": 4.9212269714521505e-05, "loss": 0.0433, "step": 727 }, { "epoch": 0.5106084516920919, "grad_norm": 1.7636412382125854, "learning_rate": 4.920865473422428e-05, "loss": 0.0171, "step": 728 }, { "epoch": 0.5113098369279326, "grad_norm": 107.01025390625, "learning_rate": 4.920503161152171e-05, "loss": 2.4773, "step": 729 }, { "epoch": 0.5120112221637735, "grad_norm": 0.35363465547561646, "learning_rate": 4.9201400347632404e-05, "loss": 0.0028, "step": 730 }, { "epoch": 0.5127126073996142, "grad_norm": 0.19405458867549896, "learning_rate": 4.919776094377771e-05, "loss": 0.0015, "step": 731 }, { "epoch": 0.5134139926354551, "grad_norm": 55.00040054321289, "learning_rate": 4.919411340118173e-05, "loss": 1.9537, "step": 732 }, { "epoch": 0.5141153778712958, "grad_norm": 55.760963439941406, "learning_rate": 4.919045772107127e-05, "loss": 2.1316, "step": 733 }, { "epoch": 0.5148167631071366, "grad_norm": 55.24003601074219, "learning_rate": 4.91867939046759e-05, "loss": 2.1231, "step": 734 }, { "epoch": 0.5155181483429774, "grad_norm": 0.03399452567100525, "learning_rate": 4.918312195322792e-05, "loss": 0.0002, "step": 735 }, { "epoch": 0.5162195335788181, "grad_norm": 204.8887176513672, "learning_rate": 4.917944186796235e-05, "loss": 8.6129, "step": 736 }, { "epoch": 0.516920918814659, "grad_norm": 124.1363754272461, "learning_rate": 4.917575365011699e-05, "loss": 3.9954, "step": 737 }, { "epoch": 0.5176223040504997, "grad_norm": 0.14609888195991516, "learning_rate": 4.9172057300932306e-05, "loss": 0.0011, "step": 738 }, { "epoch": 0.5183236892863405, "grad_norm": 0.33696234226226807, "learning_rate": 4.9168352821651575e-05, "loss": 0.0027, "step": 739 }, { "epoch": 0.5190250745221813, "grad_norm": 105.28889465332031, "learning_rate": 4.916464021352075e-05, "loss": 3.0367, "step": 740 }, { "epoch": 0.5197264597580221, "grad_norm": 1.017066240310669, "learning_rate": 4.916091947778855e-05, "loss": 0.0088, "step": 741 }, { "epoch": 0.5204278449938629, "grad_norm": 1.143968939781189, "learning_rate": 4.91571906157064e-05, "loss": 0.0102, "step": 742 }, { "epoch": 0.5211292302297037, "grad_norm": 0.7457045912742615, "learning_rate": 4.91534536285285e-05, "loss": 0.0063, "step": 743 }, { "epoch": 0.5218306154655444, "grad_norm": 0.700978696346283, "learning_rate": 4.9149708517511743e-05, "loss": 0.0058, "step": 744 }, { "epoch": 0.5225320007013853, "grad_norm": 0.38463106751441956, "learning_rate": 4.914595528391578e-05, "loss": 0.0031, "step": 745 }, { "epoch": 0.523233385937226, "grad_norm": 0.14777745306491852, "learning_rate": 4.914219392900297e-05, "loss": 0.0011, "step": 746 }, { "epoch": 0.5239347711730669, "grad_norm": 0.07884679734706879, "learning_rate": 4.913842445403843e-05, "loss": 0.0006, "step": 747 }, { "epoch": 0.5246361564089076, "grad_norm": 61.55155563354492, "learning_rate": 4.913464686028999e-05, "loss": 1.992, "step": 748 }, { "epoch": 0.5253375416447483, "grad_norm": 0.027440281584858894, "learning_rate": 4.913086114902822e-05, "loss": 0.0002, "step": 749 }, { "epoch": 0.5260389268805892, "grad_norm": 55.66230010986328, "learning_rate": 4.912706732152641e-05, "loss": 2.231, "step": 750 }, { "epoch": 0.5267403121164299, "grad_norm": 54.453102111816406, "learning_rate": 4.912326537906059e-05, "loss": 2.0983, "step": 751 }, { "epoch": 0.5274416973522708, "grad_norm": 0.0637647733092308, "learning_rate": 4.911945532290951e-05, "loss": 0.0005, "step": 752 }, { "epoch": 0.5281430825881115, "grad_norm": 0.13056063652038574, "learning_rate": 4.9115637154354674e-05, "loss": 0.001, "step": 753 }, { "epoch": 0.5288444678239523, "grad_norm": 52.72209167480469, "learning_rate": 4.911181087468027e-05, "loss": 1.5666, "step": 754 }, { "epoch": 0.5295458530597931, "grad_norm": 48.385467529296875, "learning_rate": 4.9107976485173255e-05, "loss": 1.2941, "step": 755 }, { "epoch": 0.5302472382956339, "grad_norm": 0.5080452561378479, "learning_rate": 4.91041339871233e-05, "loss": 0.0041, "step": 756 }, { "epoch": 0.5309486235314747, "grad_norm": 80.44468688964844, "learning_rate": 4.910028338182279e-05, "loss": 1.6558, "step": 757 }, { "epoch": 0.5316500087673155, "grad_norm": 1.59328031539917, "learning_rate": 4.909642467056685e-05, "loss": 0.0141, "step": 758 }, { "epoch": 0.5323513940031562, "grad_norm": 2.678886651992798, "learning_rate": 4.9092557854653344e-05, "loss": 0.0255, "step": 759 }, { "epoch": 0.533052779238997, "grad_norm": 2.742159366607666, "learning_rate": 4.908868293538282e-05, "loss": 0.0271, "step": 760 }, { "epoch": 0.5337541644748378, "grad_norm": 2.275885820388794, "learning_rate": 4.908479991405861e-05, "loss": 0.0215, "step": 761 }, { "epoch": 0.5344555497106785, "grad_norm": 1.4059484004974365, "learning_rate": 4.9080908791986724e-05, "loss": 0.0126, "step": 762 }, { "epoch": 0.5351569349465194, "grad_norm": 0.7012129426002502, "learning_rate": 4.907700957047592e-05, "loss": 0.0058, "step": 763 }, { "epoch": 0.5358583201823601, "grad_norm": 0.2713419497013092, "learning_rate": 4.9073102250837655e-05, "loss": 0.0021, "step": 764 }, { "epoch": 0.536559705418201, "grad_norm": 0.13228341937065125, "learning_rate": 4.9069186834386145e-05, "loss": 0.001, "step": 765 }, { "epoch": 0.5372610906540417, "grad_norm": 0.06449166685342789, "learning_rate": 4.90652633224383e-05, "loss": 0.0005, "step": 766 }, { "epoch": 0.5379624758898826, "grad_norm": 0.017581632360816002, "learning_rate": 4.9061331716313783e-05, "loss": 0.0001, "step": 767 }, { "epoch": 0.5386638611257233, "grad_norm": 0.016885504126548767, "learning_rate": 4.905739201733495e-05, "loss": 0.0001, "step": 768 }, { "epoch": 0.5393652463615641, "grad_norm": 0.005497175268828869, "learning_rate": 4.9053444226826886e-05, "loss": 0.0, "step": 769 }, { "epoch": 0.5400666315974049, "grad_norm": 58.73657989501953, "learning_rate": 4.90494883461174e-05, "loss": 3.1923, "step": 770 }, { "epoch": 0.5407680168332457, "grad_norm": 0.0022041036281734705, "learning_rate": 4.9045524376537034e-05, "loss": 0.0, "step": 771 }, { "epoch": 0.5414694020690864, "grad_norm": 52.654136657714844, "learning_rate": 4.9041552319419035e-05, "loss": 3.097, "step": 772 }, { "epoch": 0.5421707873049272, "grad_norm": 0.0026611709035933018, "learning_rate": 4.903757217609938e-05, "loss": 0.0, "step": 773 }, { "epoch": 0.542872172540768, "grad_norm": 0.0028983913362026215, "learning_rate": 4.9033583947916765e-05, "loss": 0.0, "step": 774 }, { "epoch": 0.5435735577766088, "grad_norm": 105.84739685058594, "learning_rate": 4.902958763621258e-05, "loss": 5.9716, "step": 775 }, { "epoch": 0.5442749430124496, "grad_norm": 54.44437789916992, "learning_rate": 4.902558324233098e-05, "loss": 2.7511, "step": 776 }, { "epoch": 0.5449763282482903, "grad_norm": 0.009829580783843994, "learning_rate": 4.9021570767618794e-05, "loss": 0.0001, "step": 777 }, { "epoch": 0.5456777134841312, "grad_norm": 53.416908264160156, "learning_rate": 4.90175502134256e-05, "loss": 2.337, "step": 778 }, { "epoch": 0.5463790987199719, "grad_norm": 0.050978660583496094, "learning_rate": 4.9013521581103686e-05, "loss": 0.0004, "step": 779 }, { "epoch": 0.5470804839558128, "grad_norm": 0.11917990446090698, "learning_rate": 4.900948487200804e-05, "loss": 0.0009, "step": 780 }, { "epoch": 0.5477818691916535, "grad_norm": 48.541015625, "learning_rate": 4.900544008749639e-05, "loss": 1.2937, "step": 781 }, { "epoch": 0.5484832544274943, "grad_norm": 0.3613051474094391, "learning_rate": 4.900138722892917e-05, "loss": 0.003, "step": 782 }, { "epoch": 0.5491846396633351, "grad_norm": 35.976261138916016, "learning_rate": 4.899732629766953e-05, "loss": 0.7424, "step": 783 }, { "epoch": 0.5498860248991758, "grad_norm": 1.3365657329559326, "learning_rate": 4.899325729508333e-05, "loss": 0.012, "step": 784 }, { "epoch": 0.5505874101350167, "grad_norm": 2.2534306049346924, "learning_rate": 4.898918022253916e-05, "loss": 0.0228, "step": 785 }, { "epoch": 0.5512887953708574, "grad_norm": 24.77885627746582, "learning_rate": 4.898509508140829e-05, "loss": 0.4601, "step": 786 }, { "epoch": 0.5519901806066982, "grad_norm": 4.28820276260376, "learning_rate": 4.898100187306475e-05, "loss": 0.0482, "step": 787 }, { "epoch": 0.552691565842539, "grad_norm": 4.493249416351318, "learning_rate": 4.897690059888524e-05, "loss": 0.0503, "step": 788 }, { "epoch": 0.5533929510783798, "grad_norm": 3.269726037979126, "learning_rate": 4.897279126024923e-05, "loss": 0.0349, "step": 789 }, { "epoch": 0.5540943363142206, "grad_norm": 3.0403542518615723, "learning_rate": 4.896867385853883e-05, "loss": 0.0316, "step": 790 }, { "epoch": 0.5547957215500614, "grad_norm": 25.840961456298828, "learning_rate": 4.896454839513892e-05, "loss": 0.4405, "step": 791 }, { "epoch": 0.5554971067859021, "grad_norm": 72.00930786132812, "learning_rate": 4.8960414871437055e-05, "loss": 1.4578, "step": 792 }, { "epoch": 0.556198492021743, "grad_norm": 0.7836804389953613, "learning_rate": 4.895627328882352e-05, "loss": 0.0068, "step": 793 }, { "epoch": 0.5568998772575837, "grad_norm": 0.7037481069564819, "learning_rate": 4.895212364869132e-05, "loss": 0.0061, "step": 794 }, { "epoch": 0.5576012624934246, "grad_norm": 0.4775547981262207, "learning_rate": 4.894796595243614e-05, "loss": 0.004, "step": 795 }, { "epoch": 0.5583026477292653, "grad_norm": 0.4537871778011322, "learning_rate": 4.894380020145639e-05, "loss": 0.0038, "step": 796 }, { "epoch": 0.559004032965106, "grad_norm": 0.2530267536640167, "learning_rate": 4.8939626397153196e-05, "loss": 0.0021, "step": 797 }, { "epoch": 0.5597054182009469, "grad_norm": 0.08755356073379517, "learning_rate": 4.8935444540930385e-05, "loss": 0.0007, "step": 798 }, { "epoch": 0.5604068034367876, "grad_norm": 0.058386508375406265, "learning_rate": 4.8931254634194504e-05, "loss": 0.0004, "step": 799 }, { "epoch": 0.5611081886726285, "grad_norm": 0.020555956289172173, "learning_rate": 4.892705667835479e-05, "loss": 0.0002, "step": 800 }, { "epoch": 0.5618095739084692, "grad_norm": 0.023624928668141365, "learning_rate": 4.892285067482319e-05, "loss": 0.0002, "step": 801 }, { "epoch": 0.56251095914431, "grad_norm": 54.471675872802734, "learning_rate": 4.891863662501437e-05, "loss": 2.8657, "step": 802 }, { "epoch": 0.5632123443801508, "grad_norm": 0.005444019101560116, "learning_rate": 4.8914414530345684e-05, "loss": 0.0, "step": 803 }, { "epoch": 0.5639137296159916, "grad_norm": 0.008992165327072144, "learning_rate": 4.891018439223722e-05, "loss": 0.0001, "step": 804 }, { "epoch": 0.5646151148518324, "grad_norm": 0.0062455967999994755, "learning_rate": 4.890594621211174e-05, "loss": 0.0, "step": 805 }, { "epoch": 0.5653165000876732, "grad_norm": 0.006425914354622364, "learning_rate": 4.8901699991394735e-05, "loss": 0.0, "step": 806 }, { "epoch": 0.5660178853235139, "grad_norm": 0.008212440647184849, "learning_rate": 4.8897445731514384e-05, "loss": 0.0001, "step": 807 }, { "epoch": 0.5667192705593547, "grad_norm": 0.004950184375047684, "learning_rate": 4.889318343390158e-05, "loss": 0.0, "step": 808 }, { "epoch": 0.5674206557951955, "grad_norm": 0.0056209079921245575, "learning_rate": 4.8888913099989916e-05, "loss": 0.0, "step": 809 }, { "epoch": 0.5681220410310363, "grad_norm": 0.0028246166184544563, "learning_rate": 4.888463473121568e-05, "loss": 0.0, "step": 810 }, { "epoch": 0.5688234262668771, "grad_norm": 53.988868713378906, "learning_rate": 4.888034832901788e-05, "loss": 2.7359, "step": 811 }, { "epoch": 0.5695248115027178, "grad_norm": 0.004621135536581278, "learning_rate": 4.8876053894838215e-05, "loss": 0.0, "step": 812 }, { "epoch": 0.5702261967385587, "grad_norm": 0.011877781711518764, "learning_rate": 4.887175143012108e-05, "loss": 0.0001, "step": 813 }, { "epoch": 0.5709275819743994, "grad_norm": 65.13773345947266, "learning_rate": 4.886744093631358e-05, "loss": 3.5609, "step": 814 }, { "epoch": 0.5716289672102403, "grad_norm": 161.86300659179688, "learning_rate": 4.886312241486554e-05, "loss": 7.2583, "step": 815 }, { "epoch": 0.572330352446081, "grad_norm": 0.032078277319669724, "learning_rate": 4.8858795867229425e-05, "loss": 0.0002, "step": 816 }, { "epoch": 0.5730317376819218, "grad_norm": 113.32162475585938, "learning_rate": 4.8854461294860454e-05, "loss": 4.1658, "step": 817 }, { "epoch": 0.5737331229177626, "grad_norm": 0.17656919360160828, "learning_rate": 4.885011869921654e-05, "loss": 0.0014, "step": 818 }, { "epoch": 0.5744345081536034, "grad_norm": 0.2188340574502945, "learning_rate": 4.884576808175828e-05, "loss": 0.0018, "step": 819 }, { "epoch": 0.5751358933894442, "grad_norm": 0.38712555170059204, "learning_rate": 4.884140944394896e-05, "loss": 0.0033, "step": 820 }, { "epoch": 0.5758372786252849, "grad_norm": 0.7210093140602112, "learning_rate": 4.883704278725458e-05, "loss": 0.0062, "step": 821 }, { "epoch": 0.5765386638611257, "grad_norm": 0.7694672346115112, "learning_rate": 4.883266811314385e-05, "loss": 0.0069, "step": 822 }, { "epoch": 0.5772400490969665, "grad_norm": 0.6353862881660461, "learning_rate": 4.8828285423088146e-05, "loss": 0.0056, "step": 823 }, { "epoch": 0.5779414343328073, "grad_norm": 87.86915588378906, "learning_rate": 4.8823894718561545e-05, "loss": 2.1381, "step": 824 }, { "epoch": 0.578642819568648, "grad_norm": 0.42429307103157043, "learning_rate": 4.8819496001040846e-05, "loss": 0.0036, "step": 825 }, { "epoch": 0.5793442048044889, "grad_norm": 0.40686339139938354, "learning_rate": 4.881508927200552e-05, "loss": 0.0034, "step": 826 }, { "epoch": 0.5800455900403296, "grad_norm": 0.27934542298316956, "learning_rate": 4.881067453293774e-05, "loss": 0.0022, "step": 827 }, { "epoch": 0.5807469752761705, "grad_norm": 44.31092834472656, "learning_rate": 4.8806251785322355e-05, "loss": 1.1003, "step": 828 }, { "epoch": 0.5814483605120112, "grad_norm": 0.17587609589099884, "learning_rate": 4.880182103064695e-05, "loss": 0.0014, "step": 829 }, { "epoch": 0.582149745747852, "grad_norm": 0.2651273012161255, "learning_rate": 4.879738227040176e-05, "loss": 0.0021, "step": 830 }, { "epoch": 0.5828511309836928, "grad_norm": 149.13035583496094, "learning_rate": 4.879293550607974e-05, "loss": 2.9684, "step": 831 }, { "epoch": 0.5835525162195335, "grad_norm": 0.17069180309772491, "learning_rate": 4.878848073917651e-05, "loss": 0.0014, "step": 832 }, { "epoch": 0.5842539014553744, "grad_norm": 0.24799254536628723, "learning_rate": 4.878401797119042e-05, "loss": 0.002, "step": 833 }, { "epoch": 0.5849552866912151, "grad_norm": 0.15171018242835999, "learning_rate": 4.8779547203622477e-05, "loss": 0.0012, "step": 834 }, { "epoch": 0.585656671927056, "grad_norm": 45.8453254699707, "learning_rate": 4.877506843797639e-05, "loss": 1.2116, "step": 835 }, { "epoch": 0.5863580571628967, "grad_norm": 70.00684356689453, "learning_rate": 4.877058167575857e-05, "loss": 2.1496, "step": 836 }, { "epoch": 0.5870594423987375, "grad_norm": 43.86625671386719, "learning_rate": 4.876608691847809e-05, "loss": 1.077, "step": 837 }, { "epoch": 0.5877608276345783, "grad_norm": 39.75180435180664, "learning_rate": 4.876158416764675e-05, "loss": 0.8959, "step": 838 }, { "epoch": 0.5884622128704191, "grad_norm": 78.85111999511719, "learning_rate": 4.8757073424779e-05, "loss": 1.609, "step": 839 }, { "epoch": 0.5891635981062598, "grad_norm": 2.562579393386841, "learning_rate": 4.8752554691392004e-05, "loss": 0.0262, "step": 840 }, { "epoch": 0.5898649833421007, "grad_norm": 3.174607276916504, "learning_rate": 4.8748027969005606e-05, "loss": 0.0331, "step": 841 }, { "epoch": 0.5905663685779414, "grad_norm": 24.42951202392578, "learning_rate": 4.874349325914232e-05, "loss": 0.494, "step": 842 }, { "epoch": 0.5912677538137823, "grad_norm": 5.76008415222168, "learning_rate": 4.87389505633274e-05, "loss": 0.069, "step": 843 }, { "epoch": 0.591969139049623, "grad_norm": 16.484464645385742, "learning_rate": 4.873439988308871e-05, "loss": 0.3677, "step": 844 }, { "epoch": 0.5926705242854637, "grad_norm": 8.441133499145508, "learning_rate": 4.872984121995686e-05, "loss": 0.1122, "step": 845 }, { "epoch": 0.5933719095213046, "grad_norm": 7.171937942504883, "learning_rate": 4.8725274575465106e-05, "loss": 0.093, "step": 846 }, { "epoch": 0.5940732947571453, "grad_norm": 7.630092620849609, "learning_rate": 4.872069995114943e-05, "loss": 0.0997, "step": 847 }, { "epoch": 0.5947746799929862, "grad_norm": 5.521340370178223, "learning_rate": 4.871611734854845e-05, "loss": 0.065, "step": 848 }, { "epoch": 0.5954760652288269, "grad_norm": 40.95110321044922, "learning_rate": 4.871152676920351e-05, "loss": 0.6631, "step": 849 }, { "epoch": 0.5961774504646677, "grad_norm": 2.602055311203003, "learning_rate": 4.8706928214658596e-05, "loss": 0.0269, "step": 850 }, { "epoch": 0.5968788357005085, "grad_norm": 30.752164840698242, "learning_rate": 4.8702321686460424e-05, "loss": 0.5944, "step": 851 }, { "epoch": 0.5975802209363493, "grad_norm": 34.97576904296875, "learning_rate": 4.8697707186158345e-05, "loss": 0.7216, "step": 852 }, { "epoch": 0.5982816061721901, "grad_norm": 1.3377373218536377, "learning_rate": 4.869308471530443e-05, "loss": 0.0126, "step": 853 }, { "epoch": 0.5989829914080309, "grad_norm": 33.73777389526367, "learning_rate": 4.868845427545341e-05, "loss": 0.6655, "step": 854 }, { "epoch": 0.5996843766438716, "grad_norm": 27.722604751586914, "learning_rate": 4.868381586816268e-05, "loss": 0.5284, "step": 855 }, { "epoch": 0.6003857618797125, "grad_norm": 39.0617561340332, "learning_rate": 4.867916949499236e-05, "loss": 0.9141, "step": 856 }, { "epoch": 0.6010871471155532, "grad_norm": 25.998207092285156, "learning_rate": 4.8674515157505205e-05, "loss": 0.4843, "step": 857 }, { "epoch": 0.601788532351394, "grad_norm": 70.33052062988281, "learning_rate": 4.8669852857266686e-05, "loss": 1.1847, "step": 858 }, { "epoch": 0.6024899175872348, "grad_norm": 2.859663724899292, "learning_rate": 4.866518259584492e-05, "loss": 0.1724, "step": 859 }, { "epoch": 0.6031913028230755, "grad_norm": 7.6815185546875, "learning_rate": 4.866050437481072e-05, "loss": 0.2053, "step": 860 }, { "epoch": 0.6038926880589164, "grad_norm": 15.574623107910156, "learning_rate": 4.8655818195737576e-05, "loss": 0.2786, "step": 861 }, { "epoch": 0.6045940732947571, "grad_norm": 17.501440048217773, "learning_rate": 4.865112406020164e-05, "loss": 0.3404, "step": 862 }, { "epoch": 0.605295458530598, "grad_norm": 18.48247718811035, "learning_rate": 4.8646421969781757e-05, "loss": 0.3676, "step": 863 }, { "epoch": 0.6059968437664387, "grad_norm": 15.630892753601074, "learning_rate": 4.864171192605944e-05, "loss": 0.2856, "step": 864 }, { "epoch": 0.6066982290022795, "grad_norm": 15.12441349029541, "learning_rate": 4.863699393061888e-05, "loss": 0.2728, "step": 865 }, { "epoch": 0.6073996142381203, "grad_norm": 10.427019119262695, "learning_rate": 4.863226798504694e-05, "loss": 0.308, "step": 866 }, { "epoch": 0.6081009994739611, "grad_norm": 10.065474510192871, "learning_rate": 4.8627534090933156e-05, "loss": 0.1617, "step": 867 }, { "epoch": 0.6088023847098019, "grad_norm": 11.481346130371094, "learning_rate": 4.862279224986973e-05, "loss": 0.2136, "step": 868 }, { "epoch": 0.6095037699456426, "grad_norm": 5.78438663482666, "learning_rate": 4.8618042463451566e-05, "loss": 0.0726, "step": 869 }, { "epoch": 0.6102051551814834, "grad_norm": 39.987937927246094, "learning_rate": 4.8613284733276206e-05, "loss": 0.8901, "step": 870 }, { "epoch": 0.6109065404173242, "grad_norm": 3.24749755859375, "learning_rate": 4.860851906094388e-05, "loss": 0.0369, "step": 871 }, { "epoch": 0.611607925653165, "grad_norm": 2.4093921184539795, "learning_rate": 4.860374544805749e-05, "loss": 0.0254, "step": 872 }, { "epoch": 0.6123093108890058, "grad_norm": 0.970287024974823, "learning_rate": 4.85989638962226e-05, "loss": 0.0092, "step": 873 }, { "epoch": 0.6130106961248466, "grad_norm": 0.4123772084712982, "learning_rate": 4.859417440704746e-05, "loss": 0.0036, "step": 874 }, { "epoch": 0.6137120813606873, "grad_norm": 0.14014503359794617, "learning_rate": 4.8589376982142977e-05, "loss": 0.0011, "step": 875 }, { "epoch": 0.6144134665965282, "grad_norm": 0.05853430926799774, "learning_rate": 4.858457162312272e-05, "loss": 0.0005, "step": 876 }, { "epoch": 0.6151148518323689, "grad_norm": 104.4763412475586, "learning_rate": 4.857975833160295e-05, "loss": 4.5228, "step": 877 }, { "epoch": 0.6158162370682098, "grad_norm": 60.22207260131836, "learning_rate": 4.857493710920258e-05, "loss": 3.2127, "step": 878 }, { "epoch": 0.6165176223040505, "grad_norm": 0.014606939628720284, "learning_rate": 4.8570107957543185e-05, "loss": 0.0001, "step": 879 }, { "epoch": 0.6172190075398913, "grad_norm": 52.49748229980469, "learning_rate": 4.856527087824902e-05, "loss": 2.5181, "step": 880 }, { "epoch": 0.6179203927757321, "grad_norm": 60.33863067626953, "learning_rate": 4.8560425872947004e-05, "loss": 3.4192, "step": 881 }, { "epoch": 0.6186217780115728, "grad_norm": 0.017534319311380386, "learning_rate": 4.8555572943266715e-05, "loss": 0.0001, "step": 882 }, { "epoch": 0.6193231632474137, "grad_norm": 0.03253171965479851, "learning_rate": 4.85507120908404e-05, "loss": 0.0002, "step": 883 }, { "epoch": 0.6200245484832544, "grad_norm": 100.64522552490234, "learning_rate": 4.8545843317302983e-05, "loss": 3.6464, "step": 884 }, { "epoch": 0.6207259337190952, "grad_norm": 52.276100158691406, "learning_rate": 4.854096662429203e-05, "loss": 1.5974, "step": 885 }, { "epoch": 0.621427318954936, "grad_norm": 0.26515480875968933, "learning_rate": 4.853608201344778e-05, "loss": 0.0022, "step": 886 }, { "epoch": 0.6221287041907768, "grad_norm": 0.4160686731338501, "learning_rate": 4.853118948641314e-05, "loss": 0.0036, "step": 887 }, { "epoch": 0.6228300894266176, "grad_norm": 40.08952331542969, "learning_rate": 4.852628904483367e-05, "loss": 0.9215, "step": 888 }, { "epoch": 0.6235314746624584, "grad_norm": 35.90172576904297, "learning_rate": 4.852138069035761e-05, "loss": 0.7604, "step": 889 }, { "epoch": 0.6242328598982991, "grad_norm": 2.5494091510772705, "learning_rate": 4.851646442463585e-05, "loss": 0.0269, "step": 890 }, { "epoch": 0.62493424513414, "grad_norm": 4.3015456199646, "learning_rate": 4.851154024932193e-05, "loss": 0.0495, "step": 891 }, { "epoch": 0.6256356303699807, "grad_norm": 3.570831537246704, "learning_rate": 4.8506608166072054e-05, "loss": 0.0385, "step": 892 }, { "epoch": 0.6263370156058214, "grad_norm": 3.575366735458374, "learning_rate": 4.8501668176545115e-05, "loss": 0.0397, "step": 893 }, { "epoch": 0.6270384008416623, "grad_norm": 2.7419331073760986, "learning_rate": 4.849672028240262e-05, "loss": 0.0288, "step": 894 }, { "epoch": 0.627739786077503, "grad_norm": 41.37546157836914, "learning_rate": 4.849176448530878e-05, "loss": 0.8202, "step": 895 }, { "epoch": 0.6284411713133439, "grad_norm": 34.27451705932617, "learning_rate": 4.8486800786930415e-05, "loss": 0.7397, "step": 896 }, { "epoch": 0.6291425565491846, "grad_norm": 1.0924946069717407, "learning_rate": 4.848182918893705e-05, "loss": 0.0104, "step": 897 }, { "epoch": 0.6298439417850255, "grad_norm": 0.9149582386016846, "learning_rate": 4.8476849693000836e-05, "loss": 0.0085, "step": 898 }, { "epoch": 0.6305453270208662, "grad_norm": 0.4877033829689026, "learning_rate": 4.8471862300796594e-05, "loss": 0.0043, "step": 899 }, { "epoch": 0.631246712256707, "grad_norm": 50.69717788696289, "learning_rate": 4.846686701400179e-05, "loss": 1.3647, "step": 900 }, { "epoch": 0.6319480974925478, "grad_norm": 45.04077911376953, "learning_rate": 4.846186383429657e-05, "loss": 1.366, "step": 901 }, { "epoch": 0.6326494827283886, "grad_norm": 47.341819763183594, "learning_rate": 4.8456852763363706e-05, "loss": 1.2603, "step": 902 }, { "epoch": 0.6333508679642293, "grad_norm": 44.485530853271484, "learning_rate": 4.8451833802888625e-05, "loss": 1.3248, "step": 903 }, { "epoch": 0.6340522532000702, "grad_norm": 0.46667349338531494, "learning_rate": 4.844680695455943e-05, "loss": 0.0041, "step": 904 }, { "epoch": 0.6347536384359109, "grad_norm": 0.7628422379493713, "learning_rate": 4.844177222006687e-05, "loss": 0.0067, "step": 905 }, { "epoch": 0.6354550236717517, "grad_norm": 0.7451947927474976, "learning_rate": 4.8436729601104327e-05, "loss": 0.0069, "step": 906 }, { "epoch": 0.6361564089075925, "grad_norm": 0.6644980907440186, "learning_rate": 4.843167909936786e-05, "loss": 0.006, "step": 907 }, { "epoch": 0.6368577941434332, "grad_norm": 41.53312683105469, "learning_rate": 4.8426620716556164e-05, "loss": 1.1098, "step": 908 }, { "epoch": 0.6375591793792741, "grad_norm": 0.5270825624465942, "learning_rate": 4.8421554454370596e-05, "loss": 0.0046, "step": 909 }, { "epoch": 0.6382605646151148, "grad_norm": 90.92555236816406, "learning_rate": 4.841648031451514e-05, "loss": 2.5683, "step": 910 }, { "epoch": 0.6389619498509557, "grad_norm": 46.597808837890625, "learning_rate": 4.841139829869646e-05, "loss": 1.4381, "step": 911 }, { "epoch": 0.6396633350867964, "grad_norm": 0.9347837567329407, "learning_rate": 4.8406308408623865e-05, "loss": 0.0083, "step": 912 }, { "epoch": 0.6403647203226372, "grad_norm": 0.8483829498291016, "learning_rate": 4.840121064600928e-05, "loss": 0.0077, "step": 913 }, { "epoch": 0.641066105558478, "grad_norm": 37.393951416015625, "learning_rate": 4.8396105012567315e-05, "loss": 0.8573, "step": 914 }, { "epoch": 0.6417674907943188, "grad_norm": 69.15557098388672, "learning_rate": 4.839099151001521e-05, "loss": 1.491, "step": 915 }, { "epoch": 0.6424688760301596, "grad_norm": 1.3047503232955933, "learning_rate": 4.838587014007284e-05, "loss": 0.0126, "step": 916 }, { "epoch": 0.6431702612660003, "grad_norm": 1.462365746498108, "learning_rate": 4.838074090446276e-05, "loss": 0.0144, "step": 917 }, { "epoch": 0.6438716465018411, "grad_norm": 1.4955204725265503, "learning_rate": 4.8375603804910154e-05, "loss": 0.0146, "step": 918 }, { "epoch": 0.6445730317376819, "grad_norm": 1.0593093633651733, "learning_rate": 4.837045884314283e-05, "loss": 0.0101, "step": 919 }, { "epoch": 0.6452744169735227, "grad_norm": 0.5433834791183472, "learning_rate": 4.836530602089127e-05, "loss": 0.0049, "step": 920 }, { "epoch": 0.6459758022093635, "grad_norm": 44.653175354003906, "learning_rate": 4.836014533988857e-05, "loss": 1.2462, "step": 921 }, { "epoch": 0.6466771874452043, "grad_norm": 0.15624181926250458, "learning_rate": 4.835497680187052e-05, "loss": 0.0013, "step": 922 }, { "epoch": 0.647378572681045, "grad_norm": 0.13333559036254883, "learning_rate": 4.8349800408575506e-05, "loss": 0.0011, "step": 923 }, { "epoch": 0.6480799579168859, "grad_norm": 0.08697400987148285, "learning_rate": 4.834461616174456e-05, "loss": 0.0007, "step": 924 }, { "epoch": 0.6487813431527266, "grad_norm": 49.62416076660156, "learning_rate": 4.8339424063121375e-05, "loss": 2.1264, "step": 925 }, { "epoch": 0.6494827283885675, "grad_norm": 47.912906646728516, "learning_rate": 4.833422411445228e-05, "loss": 1.8048, "step": 926 }, { "epoch": 0.6501841136244082, "grad_norm": 0.0816919356584549, "learning_rate": 4.832901631748622e-05, "loss": 0.0007, "step": 927 }, { "epoch": 0.650885498860249, "grad_norm": 49.09192657470703, "learning_rate": 4.8323800673974826e-05, "loss": 1.7493, "step": 928 }, { "epoch": 0.6515868840960898, "grad_norm": 98.90199279785156, "learning_rate": 4.8318577185672325e-05, "loss": 3.7461, "step": 929 }, { "epoch": 0.6522882693319305, "grad_norm": 0.22995713353157043, "learning_rate": 4.831334585433561e-05, "loss": 0.002, "step": 930 }, { "epoch": 0.6529896545677714, "grad_norm": 0.4566737413406372, "learning_rate": 4.830810668172419e-05, "loss": 0.0041, "step": 931 }, { "epoch": 0.6536910398036121, "grad_norm": 0.5328819155693054, "learning_rate": 4.830285966960023e-05, "loss": 0.0048, "step": 932 }, { "epoch": 0.6543924250394529, "grad_norm": 0.6611910462379456, "learning_rate": 4.829760481972852e-05, "loss": 0.0061, "step": 933 }, { "epoch": 0.6550938102752937, "grad_norm": 76.24475860595703, "learning_rate": 4.82923421338765e-05, "loss": 1.7814, "step": 934 }, { "epoch": 0.6557951955111345, "grad_norm": 0.7961150407791138, "learning_rate": 4.828707161381423e-05, "loss": 0.0075, "step": 935 }, { "epoch": 0.6564965807469753, "grad_norm": 33.920326232910156, "learning_rate": 4.8281793261314404e-05, "loss": 0.7362, "step": 936 }, { "epoch": 0.6571979659828161, "grad_norm": 0.9916706681251526, "learning_rate": 4.827650707815236e-05, "loss": 0.0096, "step": 937 }, { "epoch": 0.6578993512186568, "grad_norm": 0.7823458909988403, "learning_rate": 4.827121306610608e-05, "loss": 0.0074, "step": 938 }, { "epoch": 0.6586007364544977, "grad_norm": 0.8576016426086426, "learning_rate": 4.8265911226956154e-05, "loss": 0.008, "step": 939 }, { "epoch": 0.6593021216903384, "grad_norm": 0.5242276787757874, "learning_rate": 4.826060156248582e-05, "loss": 0.0047, "step": 940 }, { "epoch": 0.6600035069261792, "grad_norm": 0.31952106952667236, "learning_rate": 4.825528407448096e-05, "loss": 0.0028, "step": 941 }, { "epoch": 0.66070489216202, "grad_norm": 0.17970554530620575, "learning_rate": 4.824995876473004e-05, "loss": 0.0015, "step": 942 }, { "epoch": 0.6614062773978607, "grad_norm": 0.10021814703941345, "learning_rate": 4.824462563502422e-05, "loss": 0.0008, "step": 943 }, { "epoch": 0.6621076626337016, "grad_norm": 146.59963989257812, "learning_rate": 4.823928468715724e-05, "loss": 5.6442, "step": 944 }, { "epoch": 0.6628090478695423, "grad_norm": 0.060028087347745895, "learning_rate": 4.82339359229255e-05, "loss": 0.0005, "step": 945 }, { "epoch": 0.6635104331053832, "grad_norm": 0.05350363254547119, "learning_rate": 4.822857934412802e-05, "loss": 0.0004, "step": 946 }, { "epoch": 0.6642118183412239, "grad_norm": 197.06178283691406, "learning_rate": 4.822321495256644e-05, "loss": 8.2989, "step": 947 }, { "epoch": 0.6649132035770647, "grad_norm": 0.047274697571992874, "learning_rate": 4.821784275004503e-05, "loss": 0.0004, "step": 948 }, { "epoch": 0.6656145888129055, "grad_norm": 0.037716954946517944, "learning_rate": 4.82124627383707e-05, "loss": 0.0003, "step": 949 }, { "epoch": 0.6663159740487463, "grad_norm": 48.574623107910156, "learning_rate": 4.8207074919352976e-05, "loss": 1.7823, "step": 950 }, { "epoch": 0.667017359284587, "grad_norm": 0.06911084055900574, "learning_rate": 4.8201679294804e-05, "loss": 0.0006, "step": 951 }, { "epoch": 0.6677187445204279, "grad_norm": 0.07155507057905197, "learning_rate": 4.8196275866538574e-05, "loss": 0.0006, "step": 952 }, { "epoch": 0.6684201297562686, "grad_norm": 0.10516396909952164, "learning_rate": 4.8190864636374086e-05, "loss": 0.0009, "step": 953 }, { "epoch": 0.6691215149921094, "grad_norm": 46.50364685058594, "learning_rate": 4.818544560613056e-05, "loss": 1.4989, "step": 954 }, { "epoch": 0.6698229002279502, "grad_norm": 0.2198321670293808, "learning_rate": 4.8180018777630664e-05, "loss": 0.0019, "step": 955 }, { "epoch": 0.670524285463791, "grad_norm": 87.00233459472656, "learning_rate": 4.8174584152699665e-05, "loss": 2.3952, "step": 956 }, { "epoch": 0.6712256706996318, "grad_norm": 39.932960510253906, "learning_rate": 4.816914173316545e-05, "loss": 1.0167, "step": 957 }, { "epoch": 0.6719270559354725, "grad_norm": 0.8517984747886658, "learning_rate": 4.816369152085855e-05, "loss": 0.0082, "step": 958 }, { "epoch": 0.6726284411713134, "grad_norm": 1.0399830341339111, "learning_rate": 4.8158233517612096e-05, "loss": 0.0102, "step": 959 }, { "epoch": 0.6733298264071541, "grad_norm": 1.2987735271453857, "learning_rate": 4.815276772526186e-05, "loss": 0.013, "step": 960 }, { "epoch": 0.674031211642995, "grad_norm": 1.1722874641418457, "learning_rate": 4.8147294145646207e-05, "loss": 0.0119, "step": 961 }, { "epoch": 0.6747325968788357, "grad_norm": 30.018028259277344, "learning_rate": 4.8141812780606137e-05, "loss": 0.606, "step": 962 }, { "epoch": 0.6754339821146765, "grad_norm": 1.0533334016799927, "learning_rate": 4.813632363198527e-05, "loss": 0.0105, "step": 963 }, { "epoch": 0.6761353673505173, "grad_norm": 52.103050231933594, "learning_rate": 4.813082670162986e-05, "loss": 1.5312, "step": 964 }, { "epoch": 0.676836752586358, "grad_norm": 36.39682388305664, "learning_rate": 4.812532199138873e-05, "loss": 0.8569, "step": 965 }, { "epoch": 0.6775381378221988, "grad_norm": 84.3869400024414, "learning_rate": 4.8119809503113363e-05, "loss": 2.1346, "step": 966 }, { "epoch": 0.6782395230580396, "grad_norm": 0.8813936710357666, "learning_rate": 4.811428923865785e-05, "loss": 0.0085, "step": 967 }, { "epoch": 0.6789409082938804, "grad_norm": 1.1427128314971924, "learning_rate": 4.810876119987888e-05, "loss": 0.0113, "step": 968 }, { "epoch": 0.6796422935297212, "grad_norm": 30.400836944580078, "learning_rate": 4.810322538863578e-05, "loss": 0.6341, "step": 969 }, { "epoch": 0.680343678765562, "grad_norm": 77.0494155883789, "learning_rate": 4.8097681806790466e-05, "loss": 1.7974, "step": 970 }, { "epoch": 0.6810450640014027, "grad_norm": 29.669004440307617, "learning_rate": 4.8092130456207495e-05, "loss": 0.5951, "step": 971 }, { "epoch": 0.6817464492372436, "grad_norm": 2.0304183959960938, "learning_rate": 4.808657133875402e-05, "loss": 0.0214, "step": 972 }, { "epoch": 0.6824478344730843, "grad_norm": 22.323291778564453, "learning_rate": 4.808100445629981e-05, "loss": 0.4383, "step": 973 }, { "epoch": 0.6831492197089252, "grad_norm": 11.758098602294922, "learning_rate": 4.807542981071724e-05, "loss": 0.245, "step": 974 }, { "epoch": 0.6838506049447659, "grad_norm": 9.657792091369629, "learning_rate": 4.80698474038813e-05, "loss": 0.2276, "step": 975 }, { "epoch": 0.6845519901806068, "grad_norm": 0.9010874629020691, "learning_rate": 4.806425723766961e-05, "loss": 0.1934, "step": 976 }, { "epoch": 0.6852533754164475, "grad_norm": 13.536153793334961, "learning_rate": 4.805865931396236e-05, "loss": 0.2465, "step": 977 }, { "epoch": 0.6859547606522882, "grad_norm": 12.555607795715332, "learning_rate": 4.8053053634642384e-05, "loss": 0.2135, "step": 978 }, { "epoch": 0.6866561458881291, "grad_norm": 15.943707466125488, "learning_rate": 4.804744020159511e-05, "loss": 0.3215, "step": 979 }, { "epoch": 0.6873575311239698, "grad_norm": 9.25063419342041, "learning_rate": 4.804181901670856e-05, "loss": 0.2798, "step": 980 }, { "epoch": 0.6880589163598106, "grad_norm": 8.65345573425293, "learning_rate": 4.80361900818734e-05, "loss": 0.1354, "step": 981 }, { "epoch": 0.6887603015956514, "grad_norm": 10.924114227294922, "learning_rate": 4.803055339898288e-05, "loss": 0.1829, "step": 982 }, { "epoch": 0.6894616868314922, "grad_norm": 7.537076950073242, "learning_rate": 4.8024908969932844e-05, "loss": 0.1064, "step": 983 }, { "epoch": 0.690163072067333, "grad_norm": 43.258338928222656, "learning_rate": 4.8019256796621764e-05, "loss": 0.957, "step": 984 }, { "epoch": 0.6908644573031738, "grad_norm": 26.308055877685547, "learning_rate": 4.801359688095071e-05, "loss": 0.541, "step": 985 }, { "epoch": 0.6915658425390145, "grad_norm": 1.8512808084487915, "learning_rate": 4.800792922482333e-05, "loss": 0.0192, "step": 986 }, { "epoch": 0.6922672277748554, "grad_norm": 1.0505925416946411, "learning_rate": 4.800225383014594e-05, "loss": 0.0104, "step": 987 }, { "epoch": 0.6929686130106961, "grad_norm": 40.21651840209961, "learning_rate": 4.799657069882738e-05, "loss": 1.0056, "step": 988 }, { "epoch": 0.6936699982465369, "grad_norm": 0.17625601589679718, "learning_rate": 4.7990879832779156e-05, "loss": 0.0015, "step": 989 }, { "epoch": 0.6943713834823777, "grad_norm": 0.15104326605796814, "learning_rate": 4.7985181233915335e-05, "loss": 0.0013, "step": 990 }, { "epoch": 0.6950727687182184, "grad_norm": 0.09187794476747513, "learning_rate": 4.79794749041526e-05, "loss": 0.0008, "step": 991 }, { "epoch": 0.6957741539540593, "grad_norm": 48.85366439819336, "learning_rate": 4.797376084541024e-05, "loss": 1.9766, "step": 992 }, { "epoch": 0.6964755391899, "grad_norm": 0.04326195642352104, "learning_rate": 4.796803905961013e-05, "loss": 0.0004, "step": 993 }, { "epoch": 0.6971769244257409, "grad_norm": 0.04548042640089989, "learning_rate": 4.796230954867676e-05, "loss": 0.0004, "step": 994 }, { "epoch": 0.6978783096615816, "grad_norm": 48.174827575683594, "learning_rate": 4.79565723145372e-05, "loss": 1.9204, "step": 995 }, { "epoch": 0.6985796948974224, "grad_norm": 0.04808010160923004, "learning_rate": 4.7950827359121135e-05, "loss": 0.0004, "step": 996 }, { "epoch": 0.6992810801332632, "grad_norm": 48.49928665161133, "learning_rate": 4.794507468436084e-05, "loss": 1.8132, "step": 997 }, { "epoch": 0.699982465369104, "grad_norm": 47.405731201171875, "learning_rate": 4.793931429219117e-05, "loss": 1.6124, "step": 998 }, { "epoch": 0.7006838506049448, "grad_norm": 0.2506006062030792, "learning_rate": 4.793354618454961e-05, "loss": 0.0022, "step": 999 }, { "epoch": 0.7013852358407856, "grad_norm": 90.9857177734375, "learning_rate": 4.792777036337621e-05, "loss": 2.7266, "step": 1000 }, { "epoch": 0.7020866210766263, "grad_norm": 0.9927763938903809, "learning_rate": 4.792198683061363e-05, "loss": 0.0099, "step": 1001 }, { "epoch": 0.7027880063124671, "grad_norm": 1.5031591653823853, "learning_rate": 4.791619558820712e-05, "loss": 0.0152, "step": 1002 }, { "epoch": 0.7034893915483079, "grad_norm": 29.284543991088867, "learning_rate": 4.791039663810452e-05, "loss": 0.5838, "step": 1003 }, { "epoch": 0.7041907767841487, "grad_norm": 1.6648647785186768, "learning_rate": 4.790458998225627e-05, "loss": 0.0173, "step": 1004 }, { "epoch": 0.7048921620199895, "grad_norm": 2.1368696689605713, "learning_rate": 4.7898775622615384e-05, "loss": 0.023, "step": 1005 }, { "epoch": 0.7055935472558302, "grad_norm": 1.5531522035598755, "learning_rate": 4.789295356113749e-05, "loss": 0.0161, "step": 1006 }, { "epoch": 0.7062949324916711, "grad_norm": 1.0346095561981201, "learning_rate": 4.7887123799780795e-05, "loss": 0.0096, "step": 1007 }, { "epoch": 0.7069963177275118, "grad_norm": 0.5431362986564636, "learning_rate": 4.788128634050609e-05, "loss": 0.005, "step": 1008 }, { "epoch": 0.7076977029633527, "grad_norm": 0.19359983503818512, "learning_rate": 4.787544118527677e-05, "loss": 0.0017, "step": 1009 }, { "epoch": 0.7083990881991934, "grad_norm": 49.78437042236328, "learning_rate": 4.786958833605882e-05, "loss": 1.7876, "step": 1010 }, { "epoch": 0.7091004734350342, "grad_norm": 57.95826721191406, "learning_rate": 4.786372779482078e-05, "loss": 2.9286, "step": 1011 }, { "epoch": 0.709801858670875, "grad_norm": 48.60955047607422, "learning_rate": 4.7857859563533814e-05, "loss": 1.9456, "step": 1012 }, { "epoch": 0.7105032439067157, "grad_norm": 48.62706756591797, "learning_rate": 4.785198364417166e-05, "loss": 1.8589, "step": 1013 }, { "epoch": 0.7112046291425566, "grad_norm": 94.03495788574219, "learning_rate": 4.784610003871063e-05, "loss": 3.2697, "step": 1014 }, { "epoch": 0.7119060143783973, "grad_norm": 44.21165084838867, "learning_rate": 4.784020874912965e-05, "loss": 1.3436, "step": 1015 }, { "epoch": 0.7126073996142381, "grad_norm": 0.36148276925086975, "learning_rate": 4.7834309777410195e-05, "loss": 0.0033, "step": 1016 }, { "epoch": 0.7133087848500789, "grad_norm": 74.5863265991211, "learning_rate": 4.7828403125536354e-05, "loss": 1.8025, "step": 1017 }, { "epoch": 0.7140101700859197, "grad_norm": 43.957252502441406, "learning_rate": 4.782248879549478e-05, "loss": 1.1635, "step": 1018 }, { "epoch": 0.7147115553217604, "grad_norm": 69.43833923339844, "learning_rate": 4.781656678927471e-05, "loss": 1.2, "step": 1019 }, { "epoch": 0.7154129405576013, "grad_norm": 4.946931838989258, "learning_rate": 4.7810637108867986e-05, "loss": 0.0617, "step": 1020 }, { "epoch": 0.716114325793442, "grad_norm": 2.806675434112549, "learning_rate": 4.7804699756268987e-05, "loss": 0.1936, "step": 1021 }, { "epoch": 0.7168157110292829, "grad_norm": 10.467972755432129, "learning_rate": 4.7798754733474724e-05, "loss": 0.17, "step": 1022 }, { "epoch": 0.7175170962651236, "grad_norm": 13.816707611083984, "learning_rate": 4.7792802042484746e-05, "loss": 0.2481, "step": 1023 }, { "epoch": 0.7182184815009645, "grad_norm": 14.67568588256836, "learning_rate": 4.77868416853012e-05, "loss": 0.2728, "step": 1024 }, { "epoch": 0.7189198667368052, "grad_norm": 3.8597590923309326, "learning_rate": 4.778087366392881e-05, "loss": 0.3473, "step": 1025 }, { "epoch": 0.7196212519726459, "grad_norm": 6.207676410675049, "learning_rate": 4.777489798037489e-05, "loss": 0.276, "step": 1026 }, { "epoch": 0.7203226372084868, "grad_norm": 1.4748679399490356, "learning_rate": 4.77689146366493e-05, "loss": 0.2487, "step": 1027 }, { "epoch": 0.7210240224443275, "grad_norm": 3.2220802307128906, "learning_rate": 4.77629236347645e-05, "loss": 0.1941, "step": 1028 }, { "epoch": 0.7217254076801684, "grad_norm": 8.897248268127441, "learning_rate": 4.775692497673552e-05, "loss": 0.135, "step": 1029 }, { "epoch": 0.7224267929160091, "grad_norm": 6.12791633605957, "learning_rate": 4.7750918664579956e-05, "loss": 0.0784, "step": 1030 }, { "epoch": 0.7231281781518499, "grad_norm": 20.013917922973633, "learning_rate": 4.7744904700318006e-05, "loss": 0.3768, "step": 1031 }, { "epoch": 0.7238295633876907, "grad_norm": 2.475820541381836, "learning_rate": 4.7738883085972406e-05, "loss": 0.0274, "step": 1032 }, { "epoch": 0.7245309486235315, "grad_norm": 1.4534077644348145, "learning_rate": 4.773285382356849e-05, "loss": 0.015, "step": 1033 }, { "epoch": 0.7252323338593722, "grad_norm": 36.22199630737305, "learning_rate": 4.772681691513417e-05, "loss": 0.9046, "step": 1034 }, { "epoch": 0.7259337190952131, "grad_norm": 0.2810744643211365, "learning_rate": 4.772077236269989e-05, "loss": 0.0025, "step": 1035 }, { "epoch": 0.7266351043310538, "grad_norm": 45.44533920288086, "learning_rate": 4.771472016829871e-05, "loss": 1.3375, "step": 1036 }, { "epoch": 0.7273364895668946, "grad_norm": 0.12535035610198975, "learning_rate": 4.770866033396623e-05, "loss": 0.0011, "step": 1037 }, { "epoch": 0.7280378748027354, "grad_norm": 45.64175033569336, "learning_rate": 4.770259286174065e-05, "loss": 1.6729, "step": 1038 }, { "epoch": 0.7287392600385761, "grad_norm": 46.727394104003906, "learning_rate": 4.7696517753662696e-05, "loss": 1.7577, "step": 1039 }, { "epoch": 0.729440645274417, "grad_norm": 0.17975449562072754, "learning_rate": 4.76904350117757e-05, "loss": 0.0016, "step": 1040 }, { "epoch": 0.7301420305102577, "grad_norm": 85.70226287841797, "learning_rate": 4.7684344638125546e-05, "loss": 2.662, "step": 1041 }, { "epoch": 0.7308434157460986, "grad_norm": 39.62570571899414, "learning_rate": 4.767824663476069e-05, "loss": 1.1086, "step": 1042 }, { "epoch": 0.7315448009819393, "grad_norm": 33.90923309326172, "learning_rate": 4.767214100373214e-05, "loss": 0.8021, "step": 1043 }, { "epoch": 0.7322461862177801, "grad_norm": 23.501230239868164, "learning_rate": 4.76660277470935e-05, "loss": 0.4899, "step": 1044 }, { "epoch": 0.7329475714536209, "grad_norm": 10.125417709350586, "learning_rate": 4.76599068669009e-05, "loss": 0.2671, "step": 1045 }, { "epoch": 0.7336489566894617, "grad_norm": 3.0318775177001953, "learning_rate": 4.7653778365213076e-05, "loss": 0.2053, "step": 1046 }, { "epoch": 0.7343503419253025, "grad_norm": 7.598801612854004, "learning_rate": 4.7647642244091276e-05, "loss": 0.2937, "step": 1047 }, { "epoch": 0.7350517271611433, "grad_norm": 11.337316513061523, "learning_rate": 4.7641498505599356e-05, "loss": 0.283, "step": 1048 }, { "epoch": 0.735753112396984, "grad_norm": 2.909923553466797, "learning_rate": 4.763534715180373e-05, "loss": 0.2887, "step": 1049 }, { "epoch": 0.7364544976328248, "grad_norm": 12.51762866973877, "learning_rate": 4.762918818477334e-05, "loss": 0.2306, "step": 1050 }, { "epoch": 0.7371558828686656, "grad_norm": 10.550107955932617, "learning_rate": 4.762302160657971e-05, "loss": 0.1879, "step": 1051 }, { "epoch": 0.7378572681045064, "grad_norm": 10.876016616821289, "learning_rate": 4.7616847419296945e-05, "loss": 0.1877, "step": 1052 }, { "epoch": 0.7385586533403472, "grad_norm": 3.0670106410980225, "learning_rate": 4.761066562500167e-05, "loss": 0.2327, "step": 1053 }, { "epoch": 0.7392600385761879, "grad_norm": 5.0195770263671875, "learning_rate": 4.7604476225773084e-05, "loss": 0.0651, "step": 1054 }, { "epoch": 0.7399614238120288, "grad_norm": 2.679926872253418, "learning_rate": 4.759827922369296e-05, "loss": 0.0317, "step": 1055 }, { "epoch": 0.7406628090478695, "grad_norm": 1.752316951751709, "learning_rate": 4.7592074620845616e-05, "loss": 0.0184, "step": 1056 }, { "epoch": 0.7413641942837104, "grad_norm": 0.6544418334960938, "learning_rate": 4.758586241931791e-05, "loss": 0.0062, "step": 1057 }, { "epoch": 0.7420655795195511, "grad_norm": 0.19106708467006683, "learning_rate": 4.757964262119928e-05, "loss": 0.0017, "step": 1058 }, { "epoch": 0.742766964755392, "grad_norm": 0.0741598978638649, "learning_rate": 4.757341522858171e-05, "loss": 0.0006, "step": 1059 }, { "epoch": 0.7434683499912327, "grad_norm": 0.02510019950568676, "learning_rate": 4.7567180243559736e-05, "loss": 0.0002, "step": 1060 }, { "epoch": 0.7441697352270735, "grad_norm": 0.00887396652251482, "learning_rate": 4.7560937668230454e-05, "loss": 0.0001, "step": 1061 }, { "epoch": 0.7448711204629143, "grad_norm": 0.0027078278362751007, "learning_rate": 4.75546875046935e-05, "loss": 0.0, "step": 1062 }, { "epoch": 0.745572505698755, "grad_norm": 0.001226524356752634, "learning_rate": 4.7548429755051085e-05, "loss": 0.0, "step": 1063 }, { "epoch": 0.7462738909345958, "grad_norm": 0.0005389907164499164, "learning_rate": 4.754216442140794e-05, "loss": 0.0, "step": 1064 }, { "epoch": 0.7469752761704366, "grad_norm": 55.77132797241211, "learning_rate": 4.7535891505871375e-05, "loss": 3.9293, "step": 1065 }, { "epoch": 0.7476766614062774, "grad_norm": 103.8635025024414, "learning_rate": 4.752961101055124e-05, "loss": 7.9354, "step": 1066 }, { "epoch": 0.7483780466421182, "grad_norm": 103.83565521240234, "learning_rate": 4.7523322937559946e-05, "loss": 7.4002, "step": 1067 }, { "epoch": 0.749079431877959, "grad_norm": 0.0008341918000951409, "learning_rate": 4.751702728901242e-05, "loss": 0.0, "step": 1068 }, { "epoch": 0.7497808171137997, "grad_norm": 48.793819427490234, "learning_rate": 4.7510724067026156e-05, "loss": 3.1709, "step": 1069 }, { "epoch": 0.7504822023496406, "grad_norm": 97.3034439086914, "learning_rate": 4.750441327372122e-05, "loss": 5.7, "step": 1070 }, { "epoch": 0.7511835875854813, "grad_norm": 0.0208453219383955, "learning_rate": 4.7498094911220185e-05, "loss": 0.0002, "step": 1071 }, { "epoch": 0.7518849728213222, "grad_norm": 0.05400345101952553, "learning_rate": 4.749176898164818e-05, "loss": 0.0004, "step": 1072 }, { "epoch": 0.7525863580571629, "grad_norm": 0.0927615538239479, "learning_rate": 4.7485435487132904e-05, "loss": 0.0008, "step": 1073 }, { "epoch": 0.7532877432930036, "grad_norm": 0.19018051028251648, "learning_rate": 4.747909442980456e-05, "loss": 0.0016, "step": 1074 }, { "epoch": 0.7539891285288445, "grad_norm": 0.3820917010307312, "learning_rate": 4.7472745811795936e-05, "loss": 0.0035, "step": 1075 }, { "epoch": 0.7546905137646852, "grad_norm": 0.6652111411094666, "learning_rate": 4.746638963524235e-05, "loss": 0.0063, "step": 1076 }, { "epoch": 0.7553918990005261, "grad_norm": 91.62269592285156, "learning_rate": 4.7460025902281625e-05, "loss": 2.2848, "step": 1077 }, { "epoch": 0.7560932842363668, "grad_norm": 0.6401788592338562, "learning_rate": 4.7453654615054174e-05, "loss": 0.006, "step": 1078 }, { "epoch": 0.7567946694722076, "grad_norm": 0.9835159182548523, "learning_rate": 4.744727577570295e-05, "loss": 0.0096, "step": 1079 }, { "epoch": 0.7574960547080484, "grad_norm": 0.665366530418396, "learning_rate": 4.74408893863734e-05, "loss": 0.0064, "step": 1080 }, { "epoch": 0.7581974399438892, "grad_norm": 0.3792603313922882, "learning_rate": 4.743449544921356e-05, "loss": 0.0034, "step": 1081 }, { "epoch": 0.75889882517973, "grad_norm": 0.18616673350334167, "learning_rate": 4.742809396637397e-05, "loss": 0.0016, "step": 1082 }, { "epoch": 0.7596002104155708, "grad_norm": 0.12753531336784363, "learning_rate": 4.7421684940007725e-05, "loss": 0.0011, "step": 1083 }, { "epoch": 0.7603015956514115, "grad_norm": 92.24988555908203, "learning_rate": 4.741526837227047e-05, "loss": 3.521, "step": 1084 }, { "epoch": 0.7610029808872524, "grad_norm": 48.01845169067383, "learning_rate": 4.740884426532035e-05, "loss": 1.9461, "step": 1085 }, { "epoch": 0.7617043661230931, "grad_norm": 0.07205016911029816, "learning_rate": 4.740241262131808e-05, "loss": 0.0006, "step": 1086 }, { "epoch": 0.7624057513589338, "grad_norm": 0.07788128405809402, "learning_rate": 4.7395973442426895e-05, "loss": 0.0007, "step": 1087 }, { "epoch": 0.7631071365947747, "grad_norm": 94.46582794189453, "learning_rate": 4.738952673081256e-05, "loss": 3.4385, "step": 1088 }, { "epoch": 0.7638085218306154, "grad_norm": 0.17950232326984406, "learning_rate": 4.738307248864339e-05, "loss": 0.0016, "step": 1089 }, { "epoch": 0.7645099070664563, "grad_norm": 93.51290130615234, "learning_rate": 4.737661071809021e-05, "loss": 3.3519, "step": 1090 }, { "epoch": 0.765211292302297, "grad_norm": 92.91374206542969, "learning_rate": 4.73701414213264e-05, "loss": 2.7943, "step": 1091 }, { "epoch": 0.7659126775381379, "grad_norm": 0.5978343486785889, "learning_rate": 4.736366460052786e-05, "loss": 0.0056, "step": 1092 }, { "epoch": 0.7666140627739786, "grad_norm": 30.616474151611328, "learning_rate": 4.7357180257873014e-05, "loss": 0.6588, "step": 1093 }, { "epoch": 0.7673154480098194, "grad_norm": 1.5538511276245117, "learning_rate": 4.7350688395542825e-05, "loss": 0.0162, "step": 1094 }, { "epoch": 0.7680168332456602, "grad_norm": 75.00251007080078, "learning_rate": 4.734418901572079e-05, "loss": 1.4034, "step": 1095 }, { "epoch": 0.768718218481501, "grad_norm": 3.7650375366210938, "learning_rate": 4.733768212059293e-05, "loss": 0.0454, "step": 1096 }, { "epoch": 0.7694196037173417, "grad_norm": 5.952811241149902, "learning_rate": 4.733116771234778e-05, "loss": 0.0766, "step": 1097 }, { "epoch": 0.7701209889531825, "grad_norm": 15.600150108337402, "learning_rate": 4.7324645793176425e-05, "loss": 0.3215, "step": 1098 }, { "epoch": 0.7708223741890233, "grad_norm": 9.213759422302246, "learning_rate": 4.731811636527246e-05, "loss": 0.2548, "step": 1099 }, { "epoch": 0.7715237594248641, "grad_norm": 7.49055814743042, "learning_rate": 4.731157943083202e-05, "loss": 0.1, "step": 1100 }, { "epoch": 0.7722251446607049, "grad_norm": 7.580368518829346, "learning_rate": 4.730503499205373e-05, "loss": 0.1067, "step": 1101 }, { "epoch": 0.7729265298965456, "grad_norm": 6.691344261169434, "learning_rate": 4.7298483051138796e-05, "loss": 0.0911, "step": 1102 }, { "epoch": 0.7736279151323865, "grad_norm": 39.10125732421875, "learning_rate": 4.72919236102909e-05, "loss": 0.7558, "step": 1103 }, { "epoch": 0.7743293003682272, "grad_norm": 4.285506725311279, "learning_rate": 4.728535667171626e-05, "loss": 0.051, "step": 1104 }, { "epoch": 0.7750306856040681, "grad_norm": 2.88086199760437, "learning_rate": 4.727878223762363e-05, "loss": 0.0324, "step": 1105 }, { "epoch": 0.7757320708399088, "grad_norm": 1.6831408739089966, "learning_rate": 4.7272200310224256e-05, "loss": 0.0172, "step": 1106 }, { "epoch": 0.7764334560757497, "grad_norm": 37.848419189453125, "learning_rate": 4.726561089173194e-05, "loss": 0.934, "step": 1107 }, { "epoch": 0.7771348413115904, "grad_norm": 38.1249885559082, "learning_rate": 4.725901398436297e-05, "loss": 0.9811, "step": 1108 }, { "epoch": 0.7778362265474312, "grad_norm": 51.93446731567383, "learning_rate": 4.725240959033618e-05, "loss": 1.5657, "step": 1109 }, { "epoch": 0.778537611783272, "grad_norm": 0.33258679509162903, "learning_rate": 4.7245797711872895e-05, "loss": 0.003, "step": 1110 }, { "epoch": 0.7792389970191127, "grad_norm": 61.720924377441406, "learning_rate": 4.723917835119699e-05, "loss": 1.6662, "step": 1111 }, { "epoch": 0.7799403822549535, "grad_norm": 37.942874908447266, "learning_rate": 4.723255151053483e-05, "loss": 1.0046, "step": 1112 }, { "epoch": 0.7806417674907943, "grad_norm": 0.8135581016540527, "learning_rate": 4.7225917192115297e-05, "loss": 0.0076, "step": 1113 }, { "epoch": 0.7813431527266351, "grad_norm": 36.66408920288086, "learning_rate": 4.721927539816981e-05, "loss": 0.9099, "step": 1114 }, { "epoch": 0.7820445379624759, "grad_norm": 1.1302094459533691, "learning_rate": 4.721262613093228e-05, "loss": 0.0115, "step": 1115 }, { "epoch": 0.7827459231983167, "grad_norm": 1.4476618766784668, "learning_rate": 4.720596939263914e-05, "loss": 0.0146, "step": 1116 }, { "epoch": 0.7834473084341574, "grad_norm": 1.0591875314712524, "learning_rate": 4.719930518552934e-05, "loss": 0.0106, "step": 1117 }, { "epoch": 0.7841486936699983, "grad_norm": 0.7351242899894714, "learning_rate": 4.7192633511844334e-05, "loss": 0.0069, "step": 1118 }, { "epoch": 0.784850078905839, "grad_norm": 0.43580272793769836, "learning_rate": 4.71859543738281e-05, "loss": 0.004, "step": 1119 }, { "epoch": 0.7855514641416799, "grad_norm": 0.256906658411026, "learning_rate": 4.71792677737271e-05, "loss": 0.0022, "step": 1120 }, { "epoch": 0.7862528493775206, "grad_norm": 0.13365890085697174, "learning_rate": 4.717257371379034e-05, "loss": 0.0012, "step": 1121 }, { "epoch": 0.7869542346133613, "grad_norm": 0.07990420609712601, "learning_rate": 4.71658721962693e-05, "loss": 0.0007, "step": 1122 }, { "epoch": 0.7876556198492022, "grad_norm": 0.0479491651058197, "learning_rate": 4.715916322341801e-05, "loss": 0.0004, "step": 1123 }, { "epoch": 0.7883570050850429, "grad_norm": 47.28627014160156, "learning_rate": 4.7152446797492963e-05, "loss": 2.0982, "step": 1124 }, { "epoch": 0.7890583903208838, "grad_norm": 96.63157653808594, "learning_rate": 4.71457229207532e-05, "loss": 4.5257, "step": 1125 }, { "epoch": 0.7897597755567245, "grad_norm": 47.81488800048828, "learning_rate": 4.713899159546024e-05, "loss": 2.3268, "step": 1126 }, { "epoch": 0.7904611607925653, "grad_norm": 47.34000015258789, "learning_rate": 4.713225282387811e-05, "loss": 2.1484, "step": 1127 }, { "epoch": 0.7911625460284061, "grad_norm": 46.76436996459961, "learning_rate": 4.7125506608273344e-05, "loss": 1.8822, "step": 1128 }, { "epoch": 0.7918639312642469, "grad_norm": 0.12850354611873627, "learning_rate": 4.711875295091499e-05, "loss": 0.0011, "step": 1129 }, { "epoch": 0.7925653165000877, "grad_norm": 126.00606536865234, "learning_rate": 4.71119918540746e-05, "loss": 3.7093, "step": 1130 }, { "epoch": 0.7932667017359285, "grad_norm": 0.4526232182979584, "learning_rate": 4.7105223320026205e-05, "loss": 0.0042, "step": 1131 }, { "epoch": 0.7939680869717692, "grad_norm": 33.449928283691406, "learning_rate": 4.709844735104636e-05, "loss": 0.7578, "step": 1132 }, { "epoch": 0.7946694722076101, "grad_norm": 1.6459629535675049, "learning_rate": 4.7091663949414105e-05, "loss": 0.0174, "step": 1133 }, { "epoch": 0.7953708574434508, "grad_norm": 34.26624298095703, "learning_rate": 4.7084873117411e-05, "loss": 0.7867, "step": 1134 }, { "epoch": 0.7960722426792916, "grad_norm": 3.238154888153076, "learning_rate": 4.7078074857321085e-05, "loss": 0.0382, "step": 1135 }, { "epoch": 0.7967736279151324, "grad_norm": 3.573115348815918, "learning_rate": 4.70712691714309e-05, "loss": 0.0431, "step": 1136 }, { "epoch": 0.7974750131509731, "grad_norm": 4.002458095550537, "learning_rate": 4.7064456062029506e-05, "loss": 0.0488, "step": 1137 }, { "epoch": 0.798176398386814, "grad_norm": 3.4627814292907715, "learning_rate": 4.7057635531408424e-05, "loss": 0.0394, "step": 1138 }, { "epoch": 0.7988777836226547, "grad_norm": 18.19850730895996, "learning_rate": 4.70508075818617e-05, "loss": 0.3679, "step": 1139 }, { "epoch": 0.7995791688584956, "grad_norm": 2.7357804775238037, "learning_rate": 4.704397221568586e-05, "loss": 0.0312, "step": 1140 }, { "epoch": 0.8002805540943363, "grad_norm": 51.47498321533203, "learning_rate": 4.703712943517993e-05, "loss": 1.0259, "step": 1141 }, { "epoch": 0.8009819393301771, "grad_norm": 1.7935067415237427, "learning_rate": 4.7030279242645434e-05, "loss": 0.0192, "step": 1142 }, { "epoch": 0.8016833245660179, "grad_norm": 28.79102325439453, "learning_rate": 4.702342164038639e-05, "loss": 0.6393, "step": 1143 }, { "epoch": 0.8023847098018587, "grad_norm": 1.230648398399353, "learning_rate": 4.7016556630709284e-05, "loss": 0.0126, "step": 1144 }, { "epoch": 0.8030860950376995, "grad_norm": 0.9287470579147339, "learning_rate": 4.700968421592313e-05, "loss": 0.0094, "step": 1145 }, { "epoch": 0.8037874802735402, "grad_norm": 62.944549560546875, "learning_rate": 4.700280439833941e-05, "loss": 1.3975, "step": 1146 }, { "epoch": 0.804488865509381, "grad_norm": 44.87819290161133, "learning_rate": 4.69959171802721e-05, "loss": 1.0857, "step": 1147 }, { "epoch": 0.8051902507452218, "grad_norm": 0.5283865928649902, "learning_rate": 4.6989022564037654e-05, "loss": 0.0049, "step": 1148 }, { "epoch": 0.8058916359810626, "grad_norm": 0.5992218852043152, "learning_rate": 4.698212055195505e-05, "loss": 0.0058, "step": 1149 }, { "epoch": 0.8065930212169033, "grad_norm": 0.5411283373832703, "learning_rate": 4.697521114634571e-05, "loss": 0.0052, "step": 1150 }, { "epoch": 0.8072944064527442, "grad_norm": 75.1710433959961, "learning_rate": 4.696829434953357e-05, "loss": 1.9684, "step": 1151 }, { "epoch": 0.8079957916885849, "grad_norm": 0.3737725615501404, "learning_rate": 4.6961370163845044e-05, "loss": 0.0035, "step": 1152 }, { "epoch": 0.8086971769244258, "grad_norm": 48.369232177734375, "learning_rate": 4.6954438591609027e-05, "loss": 1.2948, "step": 1153 }, { "epoch": 0.8093985621602665, "grad_norm": 0.4170537292957306, "learning_rate": 4.694749963515691e-05, "loss": 0.0036, "step": 1154 }, { "epoch": 0.8100999473961074, "grad_norm": 0.3476961553096771, "learning_rate": 4.6940553296822565e-05, "loss": 0.0031, "step": 1155 }, { "epoch": 0.8108013326319481, "grad_norm": 40.558204650878906, "learning_rate": 4.693359957894232e-05, "loss": 1.199, "step": 1156 }, { "epoch": 0.8115027178677889, "grad_norm": 78.08576202392578, "learning_rate": 4.692663848385503e-05, "loss": 2.1697, "step": 1157 }, { "epoch": 0.8122041031036297, "grad_norm": 0.5332819819450378, "learning_rate": 4.6919670013902015e-05, "loss": 0.0051, "step": 1158 }, { "epoch": 0.8129054883394704, "grad_norm": 0.6460719108581543, "learning_rate": 4.691269417142704e-05, "loss": 0.0063, "step": 1159 }, { "epoch": 0.8136068735753113, "grad_norm": 0.6999556422233582, "learning_rate": 4.69057109587764e-05, "loss": 0.0068, "step": 1160 }, { "epoch": 0.814308258811152, "grad_norm": 0.5912299156188965, "learning_rate": 4.6898720378298846e-05, "loss": 0.0057, "step": 1161 }, { "epoch": 0.8150096440469928, "grad_norm": 0.4620411992073059, "learning_rate": 4.6891722432345597e-05, "loss": 0.0044, "step": 1162 }, { "epoch": 0.8157110292828336, "grad_norm": 0.3710283935070038, "learning_rate": 4.6884717123270374e-05, "loss": 0.0034, "step": 1163 }, { "epoch": 0.8164124145186744, "grad_norm": 0.22289921343326569, "learning_rate": 4.687770445342936e-05, "loss": 0.002, "step": 1164 }, { "epoch": 0.8171137997545151, "grad_norm": 89.52765655517578, "learning_rate": 4.687068442518121e-05, "loss": 2.8479, "step": 1165 }, { "epoch": 0.817815184990356, "grad_norm": 0.15177088975906372, "learning_rate": 4.6863657040887065e-05, "loss": 0.0013, "step": 1166 }, { "epoch": 0.8185165702261967, "grad_norm": 0.13902179896831512, "learning_rate": 4.685662230291054e-05, "loss": 0.0012, "step": 1167 }, { "epoch": 0.8192179554620376, "grad_norm": 0.10139818489551544, "learning_rate": 4.68495802136177e-05, "loss": 0.0009, "step": 1168 }, { "epoch": 0.8199193406978783, "grad_norm": 0.08662685006856918, "learning_rate": 4.684253077537712e-05, "loss": 0.0007, "step": 1169 }, { "epoch": 0.820620725933719, "grad_norm": 0.07555234432220459, "learning_rate": 4.683547399055982e-05, "loss": 0.0006, "step": 1170 }, { "epoch": 0.8213221111695599, "grad_norm": 0.06445518881082535, "learning_rate": 4.682840986153929e-05, "loss": 0.0006, "step": 1171 }, { "epoch": 0.8220234964054006, "grad_norm": 46.72000503540039, "learning_rate": 4.68213383906915e-05, "loss": 2.1281, "step": 1172 }, { "epoch": 0.8227248816412415, "grad_norm": 46.803985595703125, "learning_rate": 4.6814259580394915e-05, "loss": 1.9621, "step": 1173 }, { "epoch": 0.8234262668770822, "grad_norm": 0.07187401503324509, "learning_rate": 4.68071734330304e-05, "loss": 0.0006, "step": 1174 }, { "epoch": 0.824127652112923, "grad_norm": 0.09642617404460907, "learning_rate": 4.680007995098136e-05, "loss": 0.0008, "step": 1175 }, { "epoch": 0.8248290373487638, "grad_norm": 0.11466558277606964, "learning_rate": 4.679297913663363e-05, "loss": 0.001, "step": 1176 }, { "epoch": 0.8255304225846046, "grad_norm": 0.11334109306335449, "learning_rate": 4.67858709923755e-05, "loss": 0.001, "step": 1177 }, { "epoch": 0.8262318078204454, "grad_norm": 0.1963043361902237, "learning_rate": 4.677875552059777e-05, "loss": 0.0017, "step": 1178 }, { "epoch": 0.8269331930562862, "grad_norm": 0.17632529139518738, "learning_rate": 4.677163272369365e-05, "loss": 0.0016, "step": 1179 }, { "epoch": 0.8276345782921269, "grad_norm": 124.80657958984375, "learning_rate": 4.6764502604058864e-05, "loss": 3.9859, "step": 1180 }, { "epoch": 0.8283359635279678, "grad_norm": 0.19632282853126526, "learning_rate": 4.675736516409156e-05, "loss": 0.0018, "step": 1181 }, { "epoch": 0.8290373487638085, "grad_norm": 43.057350158691406, "learning_rate": 4.675022040619238e-05, "loss": 1.2696, "step": 1182 }, { "epoch": 0.8297387339996493, "grad_norm": 39.46339416503906, "learning_rate": 4.674306833276439e-05, "loss": 1.093, "step": 1183 }, { "epoch": 0.8304401192354901, "grad_norm": 0.653581976890564, "learning_rate": 4.6735908946213156e-05, "loss": 0.0065, "step": 1184 }, { "epoch": 0.8311415044713308, "grad_norm": 0.8602274656295776, "learning_rate": 4.6728742248946677e-05, "loss": 0.0088, "step": 1185 }, { "epoch": 0.8318428897071717, "grad_norm": 0.9501532912254333, "learning_rate": 4.6721568243375425e-05, "loss": 0.0095, "step": 1186 }, { "epoch": 0.8325442749430124, "grad_norm": 0.8702390193939209, "learning_rate": 4.671438693191232e-05, "loss": 0.0088, "step": 1187 }, { "epoch": 0.8332456601788533, "grad_norm": 0.6245111227035522, "learning_rate": 4.6707198316972746e-05, "loss": 0.0061, "step": 1188 }, { "epoch": 0.833947045414694, "grad_norm": 0.4615590572357178, "learning_rate": 4.6700002400974544e-05, "loss": 0.0045, "step": 1189 }, { "epoch": 0.8346484306505348, "grad_norm": 0.3482006788253784, "learning_rate": 4.6692799186338e-05, "loss": 0.0033, "step": 1190 }, { "epoch": 0.8353498158863756, "grad_norm": 43.396461486816406, "learning_rate": 4.668558867548587e-05, "loss": 1.3748, "step": 1191 }, { "epoch": 0.8360512011222164, "grad_norm": 49.66410827636719, "learning_rate": 4.6678370870843357e-05, "loss": 1.7113, "step": 1192 }, { "epoch": 0.8367525863580572, "grad_norm": 0.1562856137752533, "learning_rate": 4.667114577483812e-05, "loss": 0.0014, "step": 1193 }, { "epoch": 0.8374539715938979, "grad_norm": 47.60147476196289, "learning_rate": 4.666391338990026e-05, "loss": 1.8113, "step": 1194 }, { "epoch": 0.8381553568297387, "grad_norm": 0.19547221064567566, "learning_rate": 4.665667371846234e-05, "loss": 0.0018, "step": 1195 }, { "epoch": 0.8388567420655795, "grad_norm": 0.23496605455875397, "learning_rate": 4.664942676295937e-05, "loss": 0.0022, "step": 1196 }, { "epoch": 0.8395581273014203, "grad_norm": 0.19859689474105835, "learning_rate": 4.6642172525828806e-05, "loss": 0.0018, "step": 1197 }, { "epoch": 0.840259512537261, "grad_norm": 0.21561066806316376, "learning_rate": 4.6634911009510574e-05, "loss": 0.002, "step": 1198 }, { "epoch": 0.8409608977731019, "grad_norm": 0.16780033707618713, "learning_rate": 4.662764221644702e-05, "loss": 0.0016, "step": 1199 }, { "epoch": 0.8416622830089426, "grad_norm": 0.170645073056221, "learning_rate": 4.6620366149082954e-05, "loss": 0.0015, "step": 1200 }, { "epoch": 0.8423636682447835, "grad_norm": 0.16811063885688782, "learning_rate": 4.661308280986563e-05, "loss": 0.0015, "step": 1201 }, { "epoch": 0.8430650534806242, "grad_norm": 42.41083908081055, "learning_rate": 4.6605792201244733e-05, "loss": 1.4838, "step": 1202 }, { "epoch": 0.8437664387164651, "grad_norm": 43.9367790222168, "learning_rate": 4.6598494325672426e-05, "loss": 1.4149, "step": 1203 }, { "epoch": 0.8444678239523058, "grad_norm": 0.1912125051021576, "learning_rate": 4.659118918560329e-05, "loss": 0.0018, "step": 1204 }, { "epoch": 0.8451692091881466, "grad_norm": 0.27999013662338257, "learning_rate": 4.6583876783494343e-05, "loss": 0.0025, "step": 1205 }, { "epoch": 0.8458705944239874, "grad_norm": 36.06639099121094, "learning_rate": 4.657655712180507e-05, "loss": 0.9947, "step": 1206 }, { "epoch": 0.8465719796598281, "grad_norm": 0.4413144588470459, "learning_rate": 4.6569230202997385e-05, "loss": 0.0042, "step": 1207 }, { "epoch": 0.847273364895669, "grad_norm": 0.530229926109314, "learning_rate": 4.656189602953565e-05, "loss": 0.0052, "step": 1208 }, { "epoch": 0.8479747501315097, "grad_norm": 0.6561914086341858, "learning_rate": 4.655455460388665e-05, "loss": 0.0066, "step": 1209 }, { "epoch": 0.8486761353673505, "grad_norm": 0.6459203362464905, "learning_rate": 4.654720592851962e-05, "loss": 0.0063, "step": 1210 }, { "epoch": 0.8493775206031913, "grad_norm": 37.68455505371094, "learning_rate": 4.653985000590624e-05, "loss": 1.0854, "step": 1211 }, { "epoch": 0.8500789058390321, "grad_norm": 0.4941748380661011, "learning_rate": 4.6532486838520616e-05, "loss": 0.0049, "step": 1212 }, { "epoch": 0.8507802910748729, "grad_norm": 38.626068115234375, "learning_rate": 4.6525116428839307e-05, "loss": 1.0544, "step": 1213 }, { "epoch": 0.8514816763107137, "grad_norm": 0.5088845491409302, "learning_rate": 4.651773877934129e-05, "loss": 0.005, "step": 1214 }, { "epoch": 0.8521830615465544, "grad_norm": 0.5165444016456604, "learning_rate": 4.651035389250798e-05, "loss": 0.005, "step": 1215 }, { "epoch": 0.8528844467823953, "grad_norm": 37.2906379699707, "learning_rate": 4.650296177082323e-05, "loss": 1.0227, "step": 1216 }, { "epoch": 0.853585832018236, "grad_norm": 0.5255476832389832, "learning_rate": 4.649556241677333e-05, "loss": 0.0052, "step": 1217 }, { "epoch": 0.8542872172540767, "grad_norm": 37.65691375732422, "learning_rate": 4.6488155832847006e-05, "loss": 1.1044, "step": 1218 }, { "epoch": 0.8549886024899176, "grad_norm": 35.80206298828125, "learning_rate": 4.648074202153539e-05, "loss": 0.8649, "step": 1219 }, { "epoch": 0.8556899877257583, "grad_norm": 0.8690975308418274, "learning_rate": 4.647332098533208e-05, "loss": 0.0089, "step": 1220 }, { "epoch": 0.8563913729615992, "grad_norm": 1.0327463150024414, "learning_rate": 4.646589272673308e-05, "loss": 0.0109, "step": 1221 }, { "epoch": 0.8570927581974399, "grad_norm": 32.14468765258789, "learning_rate": 4.645845724823684e-05, "loss": 0.7778, "step": 1222 }, { "epoch": 0.8577941434332808, "grad_norm": 1.0274231433868408, "learning_rate": 4.645101455234422e-05, "loss": 0.0106, "step": 1223 }, { "epoch": 0.8584955286691215, "grad_norm": 29.724233627319336, "learning_rate": 4.644356464155852e-05, "loss": 0.6951, "step": 1224 }, { "epoch": 0.8591969139049623, "grad_norm": 1.3864684104919434, "learning_rate": 4.6436107518385465e-05, "loss": 0.0151, "step": 1225 }, { "epoch": 0.8598982991408031, "grad_norm": 1.3250905275344849, "learning_rate": 4.6428643185333196e-05, "loss": 0.0144, "step": 1226 }, { "epoch": 0.8605996843766439, "grad_norm": 29.683332443237305, "learning_rate": 4.6421171644912295e-05, "loss": 0.7007, "step": 1227 }, { "epoch": 0.8613010696124846, "grad_norm": 29.98341178894043, "learning_rate": 4.6413692899635764e-05, "loss": 0.6981, "step": 1228 }, { "epoch": 0.8620024548483255, "grad_norm": 30.50311279296875, "learning_rate": 4.6406206952019014e-05, "loss": 0.7341, "step": 1229 }, { "epoch": 0.8627038400841662, "grad_norm": 1.5321385860443115, "learning_rate": 4.6398713804579896e-05, "loss": 0.0171, "step": 1230 }, { "epoch": 0.863405225320007, "grad_norm": 25.173595428466797, "learning_rate": 4.6391213459838676e-05, "loss": 0.5393, "step": 1231 }, { "epoch": 0.8641066105558478, "grad_norm": 1.9056795835494995, "learning_rate": 4.638370592031804e-05, "loss": 0.0216, "step": 1232 }, { "epoch": 0.8648079957916885, "grad_norm": 54.3871955871582, "learning_rate": 4.63761911885431e-05, "loss": 1.1726, "step": 1233 }, { "epoch": 0.8655093810275294, "grad_norm": 2.3975331783294678, "learning_rate": 4.6368669267041364e-05, "loss": 0.0284, "step": 1234 }, { "epoch": 0.8662107662633701, "grad_norm": 2.3505735397338867, "learning_rate": 4.636114015834279e-05, "loss": 0.027, "step": 1235 }, { "epoch": 0.866912151499211, "grad_norm": 19.353313446044922, "learning_rate": 4.635360386497973e-05, "loss": 0.4174, "step": 1236 }, { "epoch": 0.8676135367350517, "grad_norm": 2.5313189029693604, "learning_rate": 4.6346060389486976e-05, "loss": 0.0301, "step": 1237 }, { "epoch": 0.8683149219708926, "grad_norm": 2.249114751815796, "learning_rate": 4.633850973440171e-05, "loss": 0.0262, "step": 1238 }, { "epoch": 0.8690163072067333, "grad_norm": 1.087750792503357, "learning_rate": 4.633095190226355e-05, "loss": 0.0116, "step": 1239 }, { "epoch": 0.8697176924425741, "grad_norm": 0.7214330434799194, "learning_rate": 4.63233868956145e-05, "loss": 0.0075, "step": 1240 }, { "epoch": 0.8704190776784149, "grad_norm": 0.5183669924736023, "learning_rate": 4.6315814716999003e-05, "loss": 0.0051, "step": 1241 }, { "epoch": 0.8711204629142556, "grad_norm": 0.26819220185279846, "learning_rate": 4.630823536896392e-05, "loss": 0.0026, "step": 1242 }, { "epoch": 0.8718218481500964, "grad_norm": 42.74302291870117, "learning_rate": 4.6300648854058496e-05, "loss": 1.4745, "step": 1243 }, { "epoch": 0.8725232333859372, "grad_norm": 0.06916595250368118, "learning_rate": 4.62930551748344e-05, "loss": 0.0006, "step": 1244 }, { "epoch": 0.873224618621778, "grad_norm": 0.04199717566370964, "learning_rate": 4.6285454333845714e-05, "loss": 0.0004, "step": 1245 }, { "epoch": 0.8739260038576188, "grad_norm": 46.651554107666016, "learning_rate": 4.6277846333648926e-05, "loss": 2.1377, "step": 1246 }, { "epoch": 0.8746273890934596, "grad_norm": 0.019042061641812325, "learning_rate": 4.6270231176802936e-05, "loss": 0.0002, "step": 1247 }, { "epoch": 0.8753287743293003, "grad_norm": 43.70697784423828, "learning_rate": 4.626260886586903e-05, "loss": 1.9754, "step": 1248 }, { "epoch": 0.8760301595651412, "grad_norm": 0.03968697413802147, "learning_rate": 4.625497940341094e-05, "loss": 0.0003, "step": 1249 }, { "epoch": 0.8767315448009819, "grad_norm": 0.023654410615563393, "learning_rate": 4.624734279199476e-05, "loss": 0.0002, "step": 1250 }, { "epoch": 0.8774329300368228, "grad_norm": 0.05060436204075813, "learning_rate": 4.623969903418903e-05, "loss": 0.0004, "step": 1251 }, { "epoch": 0.8781343152726635, "grad_norm": 0.062442123889923096, "learning_rate": 4.623204813256465e-05, "loss": 0.0006, "step": 1252 }, { "epoch": 0.8788357005085043, "grad_norm": 0.05655084550380707, "learning_rate": 4.622439008969497e-05, "loss": 0.0005, "step": 1253 }, { "epoch": 0.8795370857443451, "grad_norm": 0.05976495519280434, "learning_rate": 4.621672490815569e-05, "loss": 0.0005, "step": 1254 }, { "epoch": 0.8802384709801858, "grad_norm": 0.06660816818475723, "learning_rate": 4.620905259052496e-05, "loss": 0.0006, "step": 1255 }, { "epoch": 0.8809398562160267, "grad_norm": 45.397789001464844, "learning_rate": 4.6201373139383284e-05, "loss": 1.7976, "step": 1256 }, { "epoch": 0.8816412414518674, "grad_norm": 44.1479606628418, "learning_rate": 4.619368655731362e-05, "loss": 1.7646, "step": 1257 }, { "epoch": 0.8823426266877082, "grad_norm": 0.13133244216442108, "learning_rate": 4.618599284690127e-05, "loss": 0.0012, "step": 1258 }, { "epoch": 0.883044011923549, "grad_norm": 82.86041259765625, "learning_rate": 4.6178292010733984e-05, "loss": 2.727, "step": 1259 }, { "epoch": 0.8837453971593898, "grad_norm": 36.786067962646484, "learning_rate": 4.6170584051401846e-05, "loss": 1.0486, "step": 1260 }, { "epoch": 0.8844467823952306, "grad_norm": 43.94260787963867, "learning_rate": 4.6162868971497396e-05, "loss": 1.3111, "step": 1261 }, { "epoch": 0.8851481676310714, "grad_norm": 85.85136413574219, "learning_rate": 4.615514677361554e-05, "loss": 2.2235, "step": 1262 }, { "epoch": 0.8858495528669121, "grad_norm": 35.138553619384766, "learning_rate": 4.6147417460353596e-05, "loss": 0.6498, "step": 1263 }, { "epoch": 0.886550938102753, "grad_norm": 5.97971773147583, "learning_rate": 4.613968103431124e-05, "loss": 0.0857, "step": 1264 }, { "epoch": 0.8872523233385937, "grad_norm": 7.489769458770752, "learning_rate": 4.6131937498090575e-05, "loss": 0.1154, "step": 1265 }, { "epoch": 0.8879537085744345, "grad_norm": 9.91996955871582, "learning_rate": 4.612418685429608e-05, "loss": 0.1685, "step": 1266 }, { "epoch": 0.8886550938102753, "grad_norm": 2.7866861820220947, "learning_rate": 4.611642910553463e-05, "loss": 0.2187, "step": 1267 }, { "epoch": 0.889356479046116, "grad_norm": 9.682534217834473, "learning_rate": 4.610866425441549e-05, "loss": 0.1669, "step": 1268 }, { "epoch": 0.8900578642819569, "grad_norm": 8.847615242004395, "learning_rate": 4.6100892303550305e-05, "loss": 0.1458, "step": 1269 }, { "epoch": 0.8907592495177976, "grad_norm": 6.194153308868408, "learning_rate": 4.609311325555311e-05, "loss": 0.0885, "step": 1270 }, { "epoch": 0.8914606347536385, "grad_norm": 11.383880615234375, "learning_rate": 4.6085327113040346e-05, "loss": 0.2945, "step": 1271 }, { "epoch": 0.8921620199894792, "grad_norm": 41.971961975097656, "learning_rate": 4.607753387863082e-05, "loss": 0.8504, "step": 1272 }, { "epoch": 0.89286340522532, "grad_norm": 15.673029899597168, "learning_rate": 4.6069733554945725e-05, "loss": 0.3576, "step": 1273 }, { "epoch": 0.8935647904611608, "grad_norm": 2.994936466217041, "learning_rate": 4.606192614460865e-05, "loss": 0.0371, "step": 1274 }, { "epoch": 0.8942661756970016, "grad_norm": 15.890610694885254, "learning_rate": 4.6054111650245544e-05, "loss": 0.3475, "step": 1275 }, { "epoch": 0.8949675609328424, "grad_norm": 3.3848869800567627, "learning_rate": 4.604629007448479e-05, "loss": 0.0433, "step": 1276 }, { "epoch": 0.8956689461686832, "grad_norm": 26.11835289001465, "learning_rate": 4.603846141995708e-05, "loss": 0.5389, "step": 1277 }, { "epoch": 0.8963703314045239, "grad_norm": 3.1392459869384766, "learning_rate": 4.6030625689295546e-05, "loss": 0.0383, "step": 1278 }, { "epoch": 0.8970717166403647, "grad_norm": 3.49643611907959, "learning_rate": 4.6022782885135674e-05, "loss": 0.0445, "step": 1279 }, { "epoch": 0.8977731018762055, "grad_norm": 2.069392204284668, "learning_rate": 4.601493301011534e-05, "loss": 0.0228, "step": 1280 }, { "epoch": 0.8984744871120462, "grad_norm": 1.4139840602874756, "learning_rate": 4.6007076066874786e-05, "loss": 0.016, "step": 1281 }, { "epoch": 0.8991758723478871, "grad_norm": 0.8678575158119202, "learning_rate": 4.599921205805663e-05, "loss": 0.0092, "step": 1282 }, { "epoch": 0.8998772575837278, "grad_norm": 0.3188234567642212, "learning_rate": 4.5991340986305896e-05, "loss": 0.0031, "step": 1283 }, { "epoch": 0.9005786428195687, "grad_norm": 0.13479161262512207, "learning_rate": 4.598346285426994e-05, "loss": 0.0012, "step": 1284 }, { "epoch": 0.9012800280554094, "grad_norm": 44.85018539428711, "learning_rate": 4.5975577664598535e-05, "loss": 1.8893, "step": 1285 }, { "epoch": 0.9019814132912503, "grad_norm": 0.030871881172060966, "learning_rate": 4.596768541994379e-05, "loss": 0.0003, "step": 1286 }, { "epoch": 0.902682798527091, "grad_norm": 0.025855490937829018, "learning_rate": 4.595978612296021e-05, "loss": 0.0002, "step": 1287 }, { "epoch": 0.9033841837629318, "grad_norm": 46.28804397583008, "learning_rate": 4.595187977630467e-05, "loss": 2.4086, "step": 1288 }, { "epoch": 0.9040855689987726, "grad_norm": 46.17903518676758, "learning_rate": 4.5943966382636413e-05, "loss": 2.5441, "step": 1289 }, { "epoch": 0.9047869542346134, "grad_norm": 48.56827926635742, "learning_rate": 4.593604594461704e-05, "loss": 2.5474, "step": 1290 }, { "epoch": 0.9054883394704542, "grad_norm": 47.86493682861328, "learning_rate": 4.592811846491055e-05, "loss": 2.1781, "step": 1291 }, { "epoch": 0.9061897247062949, "grad_norm": 43.444766998291016, "learning_rate": 4.592018394618328e-05, "loss": 1.8371, "step": 1292 }, { "epoch": 0.9068911099421357, "grad_norm": 46.283565521240234, "learning_rate": 4.5912242391103944e-05, "loss": 1.5648, "step": 1293 }, { "epoch": 0.9075924951779765, "grad_norm": 0.2855938971042633, "learning_rate": 4.5904293802343646e-05, "loss": 0.0028, "step": 1294 }, { "epoch": 0.9082938804138173, "grad_norm": 44.066627502441406, "learning_rate": 4.589633818257582e-05, "loss": 1.3621, "step": 1295 }, { "epoch": 0.908995265649658, "grad_norm": 1.1252837181091309, "learning_rate": 4.588837553447628e-05, "loss": 0.0122, "step": 1296 }, { "epoch": 0.9096966508854989, "grad_norm": 1.7556012868881226, "learning_rate": 4.588040586072321e-05, "loss": 0.0199, "step": 1297 }, { "epoch": 0.9103980361213396, "grad_norm": 2.0839264392852783, "learning_rate": 4.5872429163997163e-05, "loss": 0.0248, "step": 1298 }, { "epoch": 0.9110994213571805, "grad_norm": 2.153773784637451, "learning_rate": 4.5864445446981025e-05, "loss": 0.0254, "step": 1299 }, { "epoch": 0.9118008065930212, "grad_norm": 2.1063156127929688, "learning_rate": 4.5856454712360067e-05, "loss": 0.0235, "step": 1300 }, { "epoch": 0.912502191828862, "grad_norm": 1.26010000705719, "learning_rate": 4.584845696282192e-05, "loss": 0.0134, "step": 1301 }, { "epoch": 0.9132035770647028, "grad_norm": 0.7219848036766052, "learning_rate": 4.584045220105656e-05, "loss": 0.0075, "step": 1302 }, { "epoch": 0.9139049623005435, "grad_norm": 0.3715592324733734, "learning_rate": 4.583244042975634e-05, "loss": 0.0037, "step": 1303 }, { "epoch": 0.9146063475363844, "grad_norm": 0.16978369653224945, "learning_rate": 4.582442165161596e-05, "loss": 0.0017, "step": 1304 }, { "epoch": 0.9153077327722251, "grad_norm": 0.10813824832439423, "learning_rate": 4.5816395869332476e-05, "loss": 0.001, "step": 1305 }, { "epoch": 0.916009118008066, "grad_norm": 0.05026814341545105, "learning_rate": 4.5808363085605296e-05, "loss": 0.0005, "step": 1306 }, { "epoch": 0.9167105032439067, "grad_norm": 0.025502556934952736, "learning_rate": 4.58003233031362e-05, "loss": 0.0002, "step": 1307 }, { "epoch": 0.9174118884797475, "grad_norm": 45.29407501220703, "learning_rate": 4.579227652462928e-05, "loss": 2.3544, "step": 1308 }, { "epoch": 0.9181132737155883, "grad_norm": 0.012745014391839504, "learning_rate": 4.578422275279105e-05, "loss": 0.0001, "step": 1309 }, { "epoch": 0.9188146589514291, "grad_norm": 0.009426114149391651, "learning_rate": 4.577616199033033e-05, "loss": 0.0001, "step": 1310 }, { "epoch": 0.9195160441872698, "grad_norm": 0.008077832870185375, "learning_rate": 4.5768094239958285e-05, "loss": 0.0001, "step": 1311 }, { "epoch": 0.9202174294231107, "grad_norm": 0.006128115579485893, "learning_rate": 4.576001950438845e-05, "loss": 0.0001, "step": 1312 }, { "epoch": 0.9209188146589514, "grad_norm": 88.39704895019531, "learning_rate": 4.575193778633671e-05, "loss": 5.5453, "step": 1313 }, { "epoch": 0.9216201998947923, "grad_norm": 45.73112869262695, "learning_rate": 4.574384908852128e-05, "loss": 2.724, "step": 1314 }, { "epoch": 0.922321585130633, "grad_norm": 45.08832550048828, "learning_rate": 4.573575341366274e-05, "loss": 2.5745, "step": 1315 }, { "epoch": 0.9230229703664737, "grad_norm": 0.01642024703323841, "learning_rate": 4.572765076448402e-05, "loss": 0.0001, "step": 1316 }, { "epoch": 0.9237243556023146, "grad_norm": 0.0225964467972517, "learning_rate": 4.5719541143710375e-05, "loss": 0.0002, "step": 1317 }, { "epoch": 0.9244257408381553, "grad_norm": 0.03809621185064316, "learning_rate": 4.571142455406943e-05, "loss": 0.0003, "step": 1318 }, { "epoch": 0.9251271260739962, "grad_norm": 42.60374450683594, "learning_rate": 4.5703300998291124e-05, "loss": 1.805, "step": 1319 }, { "epoch": 0.9258285113098369, "grad_norm": 0.08730126917362213, "learning_rate": 4.569517047910777e-05, "loss": 0.0008, "step": 1320 }, { "epoch": 0.9265298965456777, "grad_norm": 119.11152648925781, "learning_rate": 4.5687032999254e-05, "loss": 4.0878, "step": 1321 }, { "epoch": 0.9272312817815185, "grad_norm": 0.3243617117404938, "learning_rate": 4.567888856146681e-05, "loss": 0.0031, "step": 1322 }, { "epoch": 0.9279326670173593, "grad_norm": 0.4152386784553528, "learning_rate": 4.567073716848551e-05, "loss": 0.004, "step": 1323 }, { "epoch": 0.9286340522532001, "grad_norm": 30.812305450439453, "learning_rate": 4.566257882305176e-05, "loss": 0.7747, "step": 1324 }, { "epoch": 0.9293354374890409, "grad_norm": 30.550561904907227, "learning_rate": 4.565441352790957e-05, "loss": 0.7148, "step": 1325 }, { "epoch": 0.9300368227248816, "grad_norm": 79.6876220703125, "learning_rate": 4.564624128580528e-05, "loss": 1.728, "step": 1326 }, { "epoch": 0.9307382079607224, "grad_norm": 3.2627832889556885, "learning_rate": 4.5638062099487545e-05, "loss": 0.0427, "step": 1327 }, { "epoch": 0.9314395931965632, "grad_norm": 10.727757453918457, "learning_rate": 4.5629875971707395e-05, "loss": 0.2789, "step": 1328 }, { "epoch": 0.932140978432404, "grad_norm": 5.946654319763184, "learning_rate": 4.562168290521817e-05, "loss": 0.0887, "step": 1329 }, { "epoch": 0.9328423636682448, "grad_norm": 1.3037784099578857, "learning_rate": 4.561348290277554e-05, "loss": 0.1909, "step": 1330 }, { "epoch": 0.9335437489040855, "grad_norm": 8.197032928466797, "learning_rate": 4.560527596713752e-05, "loss": 0.1388, "step": 1331 }, { "epoch": 0.9342451341399264, "grad_norm": 8.003156661987305, "learning_rate": 4.559706210106446e-05, "loss": 0.1299, "step": 1332 }, { "epoch": 0.9349465193757671, "grad_norm": 13.647798538208008, "learning_rate": 4.558884130731903e-05, "loss": 0.3574, "step": 1333 }, { "epoch": 0.935647904611608, "grad_norm": 5.755825996398926, "learning_rate": 4.558061358866623e-05, "loss": 0.0832, "step": 1334 }, { "epoch": 0.9363492898474487, "grad_norm": 21.87740135192871, "learning_rate": 4.55723789478734e-05, "loss": 0.4666, "step": 1335 }, { "epoch": 0.9370506750832895, "grad_norm": 9.851075172424316, "learning_rate": 4.5564137387710196e-05, "loss": 0.2624, "step": 1336 }, { "epoch": 0.9377520603191303, "grad_norm": 4.983445644378662, "learning_rate": 4.555588891094862e-05, "loss": 0.0688, "step": 1337 }, { "epoch": 0.9384534455549711, "grad_norm": 4.145101547241211, "learning_rate": 4.554763352036296e-05, "loss": 0.0537, "step": 1338 }, { "epoch": 0.9391548307908119, "grad_norm": 4.348505020141602, "learning_rate": 4.5539371218729894e-05, "loss": 0.058, "step": 1339 }, { "epoch": 0.9398562160266526, "grad_norm": 16.51106834411621, "learning_rate": 4.5531102008828354e-05, "loss": 0.3685, "step": 1340 }, { "epoch": 0.9405576012624934, "grad_norm": 2.958770513534546, "learning_rate": 4.5522825893439645e-05, "loss": 0.0375, "step": 1341 }, { "epoch": 0.9412589864983342, "grad_norm": 1.705260157585144, "learning_rate": 4.5514542875347376e-05, "loss": 0.0194, "step": 1342 }, { "epoch": 0.941960371734175, "grad_norm": 28.18739891052246, "learning_rate": 4.550625295733748e-05, "loss": 0.6777, "step": 1343 }, { "epoch": 0.9426617569700158, "grad_norm": 39.53862380981445, "learning_rate": 4.5497956142198214e-05, "loss": 1.1208, "step": 1344 }, { "epoch": 0.9433631422058566, "grad_norm": 1.0771280527114868, "learning_rate": 4.5489652432720145e-05, "loss": 0.0118, "step": 1345 }, { "epoch": 0.9440645274416973, "grad_norm": 96.40424346923828, "learning_rate": 4.548134183169617e-05, "loss": 2.4673, "step": 1346 }, { "epoch": 0.9447659126775382, "grad_norm": 0.5970985293388367, "learning_rate": 4.547302434192149e-05, "loss": 0.0061, "step": 1347 }, { "epoch": 0.9454672979133789, "grad_norm": 36.97736358642578, "learning_rate": 4.546469996619364e-05, "loss": 0.9705, "step": 1348 }, { "epoch": 0.9461686831492198, "grad_norm": 0.6519697308540344, "learning_rate": 4.545636870731247e-05, "loss": 0.0066, "step": 1349 }, { "epoch": 0.9468700683850605, "grad_norm": 0.4174421429634094, "learning_rate": 4.544803056808013e-05, "loss": 0.0042, "step": 1350 }, { "epoch": 0.9475714536209012, "grad_norm": 0.4273722469806671, "learning_rate": 4.5439685551301094e-05, "loss": 0.0043, "step": 1351 }, { "epoch": 0.9482728388567421, "grad_norm": 0.3083280324935913, "learning_rate": 4.543133365978214e-05, "loss": 0.003, "step": 1352 }, { "epoch": 0.9489742240925828, "grad_norm": 0.2918597161769867, "learning_rate": 4.542297489633238e-05, "loss": 0.0028, "step": 1353 }, { "epoch": 0.9496756093284237, "grad_norm": 47.360557556152344, "learning_rate": 4.541460926376322e-05, "loss": 1.7171, "step": 1354 }, { "epoch": 0.9503769945642644, "grad_norm": 81.7466049194336, "learning_rate": 4.5406236764888374e-05, "loss": 2.8326, "step": 1355 }, { "epoch": 0.9510783798001052, "grad_norm": 39.265384674072266, "learning_rate": 4.539785740252388e-05, "loss": 1.3337, "step": 1356 }, { "epoch": 0.951779765035946, "grad_norm": 0.28257325291633606, "learning_rate": 4.5389471179488074e-05, "loss": 0.0027, "step": 1357 }, { "epoch": 0.9524811502717868, "grad_norm": 41.259849548339844, "learning_rate": 4.538107809860159e-05, "loss": 1.2651, "step": 1358 }, { "epoch": 0.9531825355076275, "grad_norm": 0.5483337640762329, "learning_rate": 4.537267816268739e-05, "loss": 0.0055, "step": 1359 }, { "epoch": 0.9538839207434684, "grad_norm": 0.9649966359138489, "learning_rate": 4.536427137457072e-05, "loss": 0.0098, "step": 1360 }, { "epoch": 0.9545853059793091, "grad_norm": 28.19986915588379, "learning_rate": 4.5355857737079166e-05, "loss": 0.6184, "step": 1361 }, { "epoch": 0.95528669121515, "grad_norm": 1.310431957244873, "learning_rate": 4.534743725304257e-05, "loss": 0.0142, "step": 1362 }, { "epoch": 0.9559880764509907, "grad_norm": 1.170506477355957, "learning_rate": 4.5339009925293105e-05, "loss": 0.0125, "step": 1363 }, { "epoch": 0.9566894616868314, "grad_norm": 1.135798454284668, "learning_rate": 4.533057575666526e-05, "loss": 0.0118, "step": 1364 }, { "epoch": 0.9573908469226723, "grad_norm": 0.7438721060752869, "learning_rate": 4.532213474999578e-05, "loss": 0.0077, "step": 1365 }, { "epoch": 0.958092232158513, "grad_norm": 0.46775007247924805, "learning_rate": 4.531368690812376e-05, "loss": 0.0046, "step": 1366 }, { "epoch": 0.9587936173943539, "grad_norm": 0.2829347848892212, "learning_rate": 4.530523223389056e-05, "loss": 0.0027, "step": 1367 }, { "epoch": 0.9594950026301946, "grad_norm": 0.12059792876243591, "learning_rate": 4.529677073013984e-05, "loss": 0.0011, "step": 1368 }, { "epoch": 0.9601963878660355, "grad_norm": 41.76416015625, "learning_rate": 4.5288302399717575e-05, "loss": 1.5052, "step": 1369 }, { "epoch": 0.9608977731018762, "grad_norm": 0.05551628768444061, "learning_rate": 4.5279827245472025e-05, "loss": 0.0005, "step": 1370 }, { "epoch": 0.961599158337717, "grad_norm": 44.579280853271484, "learning_rate": 4.527134527025374e-05, "loss": 1.9268, "step": 1371 }, { "epoch": 0.9623005435735578, "grad_norm": 92.77808380126953, "learning_rate": 4.526285647691559e-05, "loss": 3.8084, "step": 1372 }, { "epoch": 0.9630019288093986, "grad_norm": 0.06515678018331528, "learning_rate": 4.5254360868312694e-05, "loss": 0.0006, "step": 1373 }, { "epoch": 0.9637033140452393, "grad_norm": 44.31393814086914, "learning_rate": 4.5245858447302506e-05, "loss": 1.8936, "step": 1374 }, { "epoch": 0.9644046992810801, "grad_norm": 42.123191833496094, "learning_rate": 4.523734921674475e-05, "loss": 1.5489, "step": 1375 }, { "epoch": 0.9651060845169209, "grad_norm": 78.27152252197266, "learning_rate": 4.522883317950144e-05, "loss": 1.4444, "step": 1376 }, { "epoch": 0.9658074697527617, "grad_norm": 0.4181966185569763, "learning_rate": 4.5220310338436885e-05, "loss": 0.0041, "step": 1377 }, { "epoch": 0.9665088549886025, "grad_norm": 61.79771423339844, "learning_rate": 4.521178069641768e-05, "loss": 1.5033, "step": 1378 }, { "epoch": 0.9672102402244432, "grad_norm": 1.026915431022644, "learning_rate": 4.520324425631271e-05, "loss": 0.011, "step": 1379 }, { "epoch": 0.9679116254602841, "grad_norm": 1.1025463342666626, "learning_rate": 4.519470102099315e-05, "loss": 0.0117, "step": 1380 }, { "epoch": 0.9686130106961248, "grad_norm": 1.5587780475616455, "learning_rate": 4.518615099333243e-05, "loss": 0.018, "step": 1381 }, { "epoch": 0.9693143959319657, "grad_norm": 1.4015796184539795, "learning_rate": 4.517759417620633e-05, "loss": 0.016, "step": 1382 }, { "epoch": 0.9700157811678064, "grad_norm": 1.1269611120224, "learning_rate": 4.516903057249284e-05, "loss": 0.0124, "step": 1383 }, { "epoch": 0.9707171664036472, "grad_norm": 64.7952651977539, "learning_rate": 4.516046018507229e-05, "loss": 1.7053, "step": 1384 }, { "epoch": 0.971418551639488, "grad_norm": 36.87698745727539, "learning_rate": 4.515188301682725e-05, "loss": 0.8805, "step": 1385 }, { "epoch": 0.9721199368753288, "grad_norm": 39.0102653503418, "learning_rate": 4.5143299070642584e-05, "loss": 1.2212, "step": 1386 }, { "epoch": 0.9728213221111696, "grad_norm": 62.35078811645508, "learning_rate": 4.5134708349405457e-05, "loss": 1.6357, "step": 1387 }, { "epoch": 0.9735227073470103, "grad_norm": 31.161195755004883, "learning_rate": 4.5126110856005285e-05, "loss": 0.7379, "step": 1388 }, { "epoch": 0.9742240925828511, "grad_norm": 1.4460724592208862, "learning_rate": 4.511750659333379e-05, "loss": 0.0166, "step": 1389 }, { "epoch": 0.9749254778186919, "grad_norm": 1.6871577501296997, "learning_rate": 4.5108895564284924e-05, "loss": 0.0201, "step": 1390 }, { "epoch": 0.9756268630545327, "grad_norm": 18.601655960083008, "learning_rate": 4.510027777175497e-05, "loss": 0.3773, "step": 1391 }, { "epoch": 0.9763282482903735, "grad_norm": 24.591615676879883, "learning_rate": 4.509165321864245e-05, "loss": 0.6012, "step": 1392 }, { "epoch": 0.9770296335262143, "grad_norm": 2.3374860286712646, "learning_rate": 4.5083021907848164e-05, "loss": 0.0286, "step": 1393 }, { "epoch": 0.977731018762055, "grad_norm": 3.1211438179016113, "learning_rate": 4.50743838422752e-05, "loss": 0.0388, "step": 1394 }, { "epoch": 0.9784324039978959, "grad_norm": 15.115840911865234, "learning_rate": 4.50657390248289e-05, "loss": 0.3451, "step": 1395 }, { "epoch": 0.9791337892337366, "grad_norm": 13.081830978393555, "learning_rate": 4.50570874584169e-05, "loss": 0.3082, "step": 1396 }, { "epoch": 0.9798351744695775, "grad_norm": 3.7496583461761475, "learning_rate": 4.5048429145949076e-05, "loss": 0.0503, "step": 1397 }, { "epoch": 0.9805365597054182, "grad_norm": 22.880462646484375, "learning_rate": 4.503976409033759e-05, "loss": 0.4432, "step": 1398 }, { "epoch": 0.9812379449412589, "grad_norm": 4.944215774536133, "learning_rate": 4.503109229449688e-05, "loss": 0.0706, "step": 1399 }, { "epoch": 0.9819393301770998, "grad_norm": 4.719092845916748, "learning_rate": 4.502241376134364e-05, "loss": 0.0667, "step": 1400 }, { "epoch": 0.9826407154129405, "grad_norm": 4.378533363342285, "learning_rate": 4.5013728493796826e-05, "loss": 0.0611, "step": 1401 }, { "epoch": 0.9833421006487814, "grad_norm": 4.234534740447998, "learning_rate": 4.5005036494777666e-05, "loss": 0.055, "step": 1402 }, { "epoch": 0.9840434858846221, "grad_norm": 2.942084789276123, "learning_rate": 4.4996337767209656e-05, "loss": 0.0382, "step": 1403 }, { "epoch": 0.9847448711204629, "grad_norm": 1.5875838994979858, "learning_rate": 4.498763231401855e-05, "loss": 0.018, "step": 1404 }, { "epoch": 0.9854462563563037, "grad_norm": 53.9989128112793, "learning_rate": 4.497892013813235e-05, "loss": 1.2552, "step": 1405 }, { "epoch": 0.9861476415921445, "grad_norm": 28.49068832397461, "learning_rate": 4.497020124248135e-05, "loss": 0.7336, "step": 1406 }, { "epoch": 0.9868490268279853, "grad_norm": 0.4417467713356018, "learning_rate": 4.496147562999807e-05, "loss": 0.0045, "step": 1407 }, { "epoch": 0.9875504120638261, "grad_norm": 104.99906921386719, "learning_rate": 4.4952743303617325e-05, "loss": 3.0756, "step": 1408 }, { "epoch": 0.9882517972996668, "grad_norm": 0.3293454647064209, "learning_rate": 4.4944004266276164e-05, "loss": 0.0033, "step": 1409 }, { "epoch": 0.9889531825355077, "grad_norm": 41.633243560791016, "learning_rate": 4.4935258520913894e-05, "loss": 1.3174, "step": 1410 }, { "epoch": 0.9896545677713484, "grad_norm": 0.23380160331726074, "learning_rate": 4.4926506070472085e-05, "loss": 0.0023, "step": 1411 }, { "epoch": 0.9903559530071891, "grad_norm": 0.44888490438461304, "learning_rate": 4.4917746917894556e-05, "loss": 0.0047, "step": 1412 }, { "epoch": 0.99105733824303, "grad_norm": 33.958248138427734, "learning_rate": 4.490898106612739e-05, "loss": 0.9667, "step": 1413 }, { "epoch": 0.9917587234788707, "grad_norm": 0.3081722855567932, "learning_rate": 4.490020851811892e-05, "loss": 0.003, "step": 1414 }, { "epoch": 0.9924601087147116, "grad_norm": 32.04133224487305, "learning_rate": 4.4891429276819726e-05, "loss": 0.9003, "step": 1415 }, { "epoch": 0.9931614939505523, "grad_norm": 43.6740608215332, "learning_rate": 4.488264334518264e-05, "loss": 0.9288, "step": 1416 }, { "epoch": 0.9938628791863932, "grad_norm": 29.450580596923828, "learning_rate": 4.487385072616275e-05, "loss": 0.7131, "step": 1417 }, { "epoch": 0.9945642644222339, "grad_norm": 1.1718950271606445, "learning_rate": 4.486505142271739e-05, "loss": 0.0137, "step": 1418 }, { "epoch": 0.9952656496580747, "grad_norm": 1.5927923917770386, "learning_rate": 4.485624543780613e-05, "loss": 0.0192, "step": 1419 }, { "epoch": 0.9959670348939155, "grad_norm": 24.835670471191406, "learning_rate": 4.4847432774390807e-05, "loss": 0.5717, "step": 1420 }, { "epoch": 0.9966684201297563, "grad_norm": 1.8400554656982422, "learning_rate": 4.4838613435435505e-05, "loss": 0.022, "step": 1421 }, { "epoch": 0.997369805365597, "grad_norm": 2.3654353618621826, "learning_rate": 4.482978742390654e-05, "loss": 0.0298, "step": 1422 }, { "epoch": 0.9980711906014378, "grad_norm": 24.44013023376465, "learning_rate": 4.482095474277247e-05, "loss": 0.5608, "step": 1423 }, { "epoch": 0.9987725758372786, "grad_norm": 2.5648183822631836, "learning_rate": 4.48121153950041e-05, "loss": 0.034, "step": 1424 }, { "epoch": 0.9994739610731194, "grad_norm": 2.0194385051727295, "learning_rate": 4.4803269383574507e-05, "loss": 0.0249, "step": 1425 } ], "logging_steps": 1, "max_steps": 1425, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.863941975950848e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }