diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,10008 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9994739610731194, + "eval_steps": 500, + "global_step": 1425, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0007013852358407856, + "grad_norm": NaN, + "learning_rate": 0.0, + "loss": 0.9295, + "step": 1 + }, + { + "epoch": 0.0014027704716815712, + "grad_norm": NaN, + "learning_rate": 0.0, + "loss": 1.432, + "step": 2 + }, + { + "epoch": 0.0021041557075223566, + "grad_norm": NaN, + "learning_rate": 0.0, + "loss": 0.3822, + "step": 3 + }, + { + "epoch": 0.0028055409433631424, + "grad_norm": NaN, + "learning_rate": 0.0, + "loss": 0.6542, + "step": 4 + }, + { + "epoch": 0.003506926179203928, + "grad_norm": 29.30962562561035, + "learning_rate": 1.7543859649122808e-07, + "loss": 0.3371, + "step": 5 + }, + { + "epoch": 0.004208311415044713, + "grad_norm": 34.5477294921875, + "learning_rate": 3.5087719298245616e-07, + "loss": 0.1946, + "step": 6 + }, + { + "epoch": 0.004909696650885499, + "grad_norm": 55.77326965332031, + "learning_rate": 5.263157894736843e-07, + "loss": 0.5682, + "step": 7 + }, + { + "epoch": 0.005611081886726285, + "grad_norm": 52.84772491455078, + "learning_rate": 7.017543859649123e-07, + "loss": 0.3959, + "step": 8 + }, + { + "epoch": 0.00631246712256707, + "grad_norm": 44.6428337097168, + "learning_rate": 8.771929824561404e-07, + "loss": 0.2899, + "step": 9 + }, + { + "epoch": 0.007013852358407856, + "grad_norm": 69.89168548583984, + "learning_rate": 1.0526315789473685e-06, + "loss": 0.6255, + "step": 10 + }, + { + "epoch": 0.007715237594248641, + "grad_norm": NaN, + "learning_rate": 1.0526315789473685e-06, + "loss": 0.4706, + "step": 11 + }, + { + "epoch": 0.008416622830089426, + "grad_norm": 65.1708755493164, + "learning_rate": 1.2280701754385965e-06, + "loss": 0.6875, + "step": 12 + }, + { + "epoch": 0.009118008065930211, + "grad_norm": 62.44850158691406, + "learning_rate": 1.4035087719298246e-06, + "loss": 0.8691, + "step": 13 + }, + { + "epoch": 0.009819393301770998, + "grad_norm": 50.52750778198242, + "learning_rate": 1.5789473684210528e-06, + "loss": 0.5682, + "step": 14 + }, + { + "epoch": 0.010520778537611783, + "grad_norm": 38.50554656982422, + "learning_rate": 1.7543859649122807e-06, + "loss": 0.4118, + "step": 15 + }, + { + "epoch": 0.01122216377345257, + "grad_norm": Infinity, + "learning_rate": 1.7543859649122807e-06, + "loss": 0.4622, + "step": 16 + }, + { + "epoch": 0.011923549009293355, + "grad_norm": 41.367347717285156, + "learning_rate": 1.929824561403509e-06, + "loss": 0.2331, + "step": 17 + }, + { + "epoch": 0.01262493424513414, + "grad_norm": 47.666603088378906, + "learning_rate": 2.105263157894737e-06, + "loss": 0.489, + "step": 18 + }, + { + "epoch": 0.013326319480974926, + "grad_norm": NaN, + "learning_rate": 2.105263157894737e-06, + "loss": 6.2172, + "step": 19 + }, + { + "epoch": 0.014027704716815711, + "grad_norm": 161.6815643310547, + "learning_rate": 2.2807017543859652e-06, + "loss": 1.9947, + "step": 20 + }, + { + "epoch": 0.014729089952656496, + "grad_norm": 14.031229019165039, + "learning_rate": 2.456140350877193e-06, + "loss": 0.0529, + "step": 21 + }, + { + "epoch": 0.015430475188497283, + "grad_norm": 51.389862060546875, + "learning_rate": 2.631578947368421e-06, + "loss": 0.3272, + "step": 22 + }, + { + "epoch": 0.016131860424338066, + "grad_norm": 16.31728744506836, + "learning_rate": 2.8070175438596493e-06, + "loss": 0.0668, + "step": 23 + }, + { + "epoch": 0.016833245660178853, + "grad_norm": 32.19120407104492, + "learning_rate": 2.9824561403508774e-06, + "loss": 0.1298, + "step": 24 + }, + { + "epoch": 0.01753463089601964, + "grad_norm": 8.127516746520996, + "learning_rate": 3.1578947368421056e-06, + "loss": 0.0245, + "step": 25 + }, + { + "epoch": 0.018236016131860423, + "grad_norm": 31.48802375793457, + "learning_rate": 3.3333333333333333e-06, + "loss": 0.1808, + "step": 26 + }, + { + "epoch": 0.01893740136770121, + "grad_norm": 174.7755584716797, + "learning_rate": 3.5087719298245615e-06, + "loss": 1.5976, + "step": 27 + }, + { + "epoch": 0.019638786603541996, + "grad_norm": 173.22021484375, + "learning_rate": 3.6842105263157892e-06, + "loss": 2.0731, + "step": 28 + }, + { + "epoch": 0.020340171839382783, + "grad_norm": 16.176433563232422, + "learning_rate": 3.859649122807018e-06, + "loss": 0.0533, + "step": 29 + }, + { + "epoch": 0.021041557075223566, + "grad_norm": 25.456438064575195, + "learning_rate": 4.035087719298246e-06, + "loss": 0.2239, + "step": 30 + }, + { + "epoch": 0.021742942311064353, + "grad_norm": 2.50178861618042, + "learning_rate": 4.210526315789474e-06, + "loss": 0.007, + "step": 31 + }, + { + "epoch": 0.02244432754690514, + "grad_norm": 8.369728088378906, + "learning_rate": 4.3859649122807014e-06, + "loss": 0.0312, + "step": 32 + }, + { + "epoch": 0.023145712782745922, + "grad_norm": 1.419299602508545, + "learning_rate": 4.5614035087719304e-06, + "loss": 0.0039, + "step": 33 + }, + { + "epoch": 0.02384709801858671, + "grad_norm": 21.602149963378906, + "learning_rate": 4.736842105263159e-06, + "loss": 0.1907, + "step": 34 + }, + { + "epoch": 0.024548483254427496, + "grad_norm": 96.2072525024414, + "learning_rate": 4.912280701754386e-06, + "loss": 0.4621, + "step": 35 + }, + { + "epoch": 0.02524986849026828, + "grad_norm": 186.85726928710938, + "learning_rate": 5.087719298245614e-06, + "loss": 3.6381, + "step": 36 + }, + { + "epoch": 0.025951253726109066, + "grad_norm": 222.0509033203125, + "learning_rate": 5.263157894736842e-06, + "loss": 2.1687, + "step": 37 + }, + { + "epoch": 0.026652638961949852, + "grad_norm": 395.810791015625, + "learning_rate": 5.43859649122807e-06, + "loss": 4.1469, + "step": 38 + }, + { + "epoch": 0.027354024197790636, + "grad_norm": 24.080060958862305, + "learning_rate": 5.6140350877192985e-06, + "loss": 0.0944, + "step": 39 + }, + { + "epoch": 0.028055409433631422, + "grad_norm": 23.396860122680664, + "learning_rate": 5.789473684210527e-06, + "loss": 0.2287, + "step": 40 + }, + { + "epoch": 0.02875679466947221, + "grad_norm": 26.664474487304688, + "learning_rate": 5.964912280701755e-06, + "loss": 0.2467, + "step": 41 + }, + { + "epoch": 0.029458179905312992, + "grad_norm": 0.3562469482421875, + "learning_rate": 6.140350877192982e-06, + "loss": 0.0005, + "step": 42 + }, + { + "epoch": 0.03015956514115378, + "grad_norm": Infinity, + "learning_rate": 6.140350877192982e-06, + "loss": 1.6967, + "step": 43 + }, + { + "epoch": 0.030860950376994566, + "grad_norm": 0.1253829002380371, + "learning_rate": 6.315789473684211e-06, + "loss": 0.0003, + "step": 44 + }, + { + "epoch": 0.03156233561283535, + "grad_norm": 18.994827270507812, + "learning_rate": 6.4912280701754385e-06, + "loss": 0.233, + "step": 45 + }, + { + "epoch": 0.03226372084867613, + "grad_norm": 34.921852111816406, + "learning_rate": 6.666666666666667e-06, + "loss": 0.3528, + "step": 46 + }, + { + "epoch": 0.03296510608451692, + "grad_norm": 202.69903564453125, + "learning_rate": 6.842105263157896e-06, + "loss": 2.7552, + "step": 47 + }, + { + "epoch": 0.033666491320357705, + "grad_norm": 359.0797424316406, + "learning_rate": 7.017543859649123e-06, + "loss": 2.112, + "step": 48 + }, + { + "epoch": 0.03436787655619849, + "grad_norm": 3.0262975692749023, + "learning_rate": 7.192982456140351e-06, + "loss": 0.0094, + "step": 49 + }, + { + "epoch": 0.03506926179203928, + "grad_norm": 504.8307189941406, + "learning_rate": 7.3684210526315784e-06, + "loss": 6.4525, + "step": 50 + }, + { + "epoch": 0.03577064702788006, + "grad_norm": 7.283820152282715, + "learning_rate": 7.5438596491228074e-06, + "loss": 0.0419, + "step": 51 + }, + { + "epoch": 0.036472032263720845, + "grad_norm": 10.193113327026367, + "learning_rate": 7.719298245614036e-06, + "loss": 0.0805, + "step": 52 + }, + { + "epoch": 0.037173417499561635, + "grad_norm": 21.247915267944336, + "learning_rate": 7.894736842105263e-06, + "loss": 0.1385, + "step": 53 + }, + { + "epoch": 0.03787480273540242, + "grad_norm": 12.750287055969238, + "learning_rate": 8.070175438596492e-06, + "loss": 0.1786, + "step": 54 + }, + { + "epoch": 0.03857618797124321, + "grad_norm": 186.3012237548828, + "learning_rate": 8.245614035087721e-06, + "loss": 2.5431, + "step": 55 + }, + { + "epoch": 0.03927757320708399, + "grad_norm": 0.6861621141433716, + "learning_rate": 8.421052631578948e-06, + "loss": 0.0019, + "step": 56 + }, + { + "epoch": 0.039978958442924775, + "grad_norm": 0.36288923025131226, + "learning_rate": 8.596491228070176e-06, + "loss": 0.0008, + "step": 57 + }, + { + "epoch": 0.040680343678765565, + "grad_norm": 0.08618122339248657, + "learning_rate": 8.771929824561403e-06, + "loss": 0.0001, + "step": 58 + }, + { + "epoch": 0.04138172891460635, + "grad_norm": 341.2626037597656, + "learning_rate": 8.947368421052632e-06, + "loss": 3.6742, + "step": 59 + }, + { + "epoch": 0.04208311415044713, + "grad_norm": 5.118091583251953, + "learning_rate": 9.122807017543861e-06, + "loss": 0.0128, + "step": 60 + }, + { + "epoch": 0.04278449938628792, + "grad_norm": 3.345242977142334, + "learning_rate": 9.298245614035088e-06, + "loss": 0.0095, + "step": 61 + }, + { + "epoch": 0.043485884622128705, + "grad_norm": 319.7175598144531, + "learning_rate": 9.473684210526317e-06, + "loss": 4.3047, + "step": 62 + }, + { + "epoch": 0.04418726985796949, + "grad_norm": 279.8736267089844, + "learning_rate": 9.649122807017545e-06, + "loss": 1.8524, + "step": 63 + }, + { + "epoch": 0.04488865509381028, + "grad_norm": 187.50588989257812, + "learning_rate": 9.824561403508772e-06, + "loss": 2.2166, + "step": 64 + }, + { + "epoch": 0.04559004032965106, + "grad_norm": 3.1667685508728027, + "learning_rate": 1e-05, + "loss": 0.0051, + "step": 65 + }, + { + "epoch": 0.046291425565491845, + "grad_norm": 347.2233581542969, + "learning_rate": 1.0175438596491228e-05, + "loss": 6.7317, + "step": 66 + }, + { + "epoch": 0.046992810801332635, + "grad_norm": 0.21984274685382843, + "learning_rate": 1.0350877192982457e-05, + "loss": 0.0006, + "step": 67 + }, + { + "epoch": 0.04769419603717342, + "grad_norm": 657.2495727539062, + "learning_rate": 1.0526315789473684e-05, + "loss": 12.4963, + "step": 68 + }, + { + "epoch": 0.0483955812730142, + "grad_norm": 180.615234375, + "learning_rate": 1.0701754385964913e-05, + "loss": 3.2259, + "step": 69 + }, + { + "epoch": 0.04909696650885499, + "grad_norm": 0.4523675739765167, + "learning_rate": 1.087719298245614e-05, + "loss": 0.0011, + "step": 70 + }, + { + "epoch": 0.049798351744695775, + "grad_norm": 241.3996124267578, + "learning_rate": 1.1052631578947368e-05, + "loss": 2.9636, + "step": 71 + }, + { + "epoch": 0.05049973698053656, + "grad_norm": 0.21977363526821136, + "learning_rate": 1.1228070175438597e-05, + "loss": 0.0005, + "step": 72 + }, + { + "epoch": 0.05120112221637735, + "grad_norm": 102.25936889648438, + "learning_rate": 1.1403508771929824e-05, + "loss": 0.4978, + "step": 73 + }, + { + "epoch": 0.05190250745221813, + "grad_norm": 24.031539916992188, + "learning_rate": 1.1578947368421053e-05, + "loss": 0.119, + "step": 74 + }, + { + "epoch": 0.052603892688058915, + "grad_norm": 187.41812133789062, + "learning_rate": 1.1754385964912282e-05, + "loss": 1.0223, + "step": 75 + }, + { + "epoch": 0.053305277923899705, + "grad_norm": 45.128421783447266, + "learning_rate": 1.192982456140351e-05, + "loss": 0.4026, + "step": 76 + }, + { + "epoch": 0.05400666315974049, + "grad_norm": 12.320211410522461, + "learning_rate": 1.2105263157894737e-05, + "loss": 0.1124, + "step": 77 + }, + { + "epoch": 0.05470804839558127, + "grad_norm": 26.743837356567383, + "learning_rate": 1.2280701754385964e-05, + "loss": 0.1488, + "step": 78 + }, + { + "epoch": 0.05540943363142206, + "grad_norm": 30.013471603393555, + "learning_rate": 1.2456140350877193e-05, + "loss": 0.3796, + "step": 79 + }, + { + "epoch": 0.056110818867262845, + "grad_norm": 38.910221099853516, + "learning_rate": 1.2631578947368422e-05, + "loss": 0.7097, + "step": 80 + }, + { + "epoch": 0.05681220410310363, + "grad_norm": 26.001049041748047, + "learning_rate": 1.2807017543859651e-05, + "loss": 0.1451, + "step": 81 + }, + { + "epoch": 0.05751358933894442, + "grad_norm": 12.577742576599121, + "learning_rate": 1.2982456140350877e-05, + "loss": 0.0548, + "step": 82 + }, + { + "epoch": 0.0582149745747852, + "grad_norm": 334.09478759765625, + "learning_rate": 1.3157894736842106e-05, + "loss": 8.8326, + "step": 83 + }, + { + "epoch": 0.058916359810625984, + "grad_norm": 33.031837463378906, + "learning_rate": 1.3333333333333333e-05, + "loss": 0.2555, + "step": 84 + }, + { + "epoch": 0.059617745046466775, + "grad_norm": 47.51133346557617, + "learning_rate": 1.3508771929824562e-05, + "loss": 0.4288, + "step": 85 + }, + { + "epoch": 0.06031913028230756, + "grad_norm": 41.160430908203125, + "learning_rate": 1.3684210526315791e-05, + "loss": 0.3075, + "step": 86 + }, + { + "epoch": 0.06102051551814834, + "grad_norm": 9.073790550231934, + "learning_rate": 1.3859649122807017e-05, + "loss": 0.045, + "step": 87 + }, + { + "epoch": 0.06172190075398913, + "grad_norm": 352.0498352050781, + "learning_rate": 1.4035087719298246e-05, + "loss": 1.3396, + "step": 88 + }, + { + "epoch": 0.062423285989829914, + "grad_norm": 36.5557746887207, + "learning_rate": 1.4210526315789475e-05, + "loss": 0.1926, + "step": 89 + }, + { + "epoch": 0.0631246712256707, + "grad_norm": 6.991324424743652, + "learning_rate": 1.4385964912280702e-05, + "loss": 0.0341, + "step": 90 + }, + { + "epoch": 0.06382605646151149, + "grad_norm": 6.0483078956604, + "learning_rate": 1.4561403508771931e-05, + "loss": 0.0219, + "step": 91 + }, + { + "epoch": 0.06452744169735226, + "grad_norm": 4.510119438171387, + "learning_rate": 1.4736842105263157e-05, + "loss": 0.0161, + "step": 92 + }, + { + "epoch": 0.06522882693319305, + "grad_norm": 0.8209487199783325, + "learning_rate": 1.4912280701754386e-05, + "loss": 0.0022, + "step": 93 + }, + { + "epoch": 0.06593021216903384, + "grad_norm": 0.0308829378336668, + "learning_rate": 1.5087719298245615e-05, + "loss": 0.0001, + "step": 94 + }, + { + "epoch": 0.06663159740487462, + "grad_norm": 171.68765258789062, + "learning_rate": 1.5263157894736842e-05, + "loss": 4.8668, + "step": 95 + }, + { + "epoch": 0.06733298264071541, + "grad_norm": 0.0008907412411645055, + "learning_rate": 1.543859649122807e-05, + "loss": 0.0, + "step": 96 + }, + { + "epoch": 0.0680343678765562, + "grad_norm": 0.0062671443447470665, + "learning_rate": 1.56140350877193e-05, + "loss": 0.0, + "step": 97 + }, + { + "epoch": 0.06873575311239698, + "grad_norm": 0.02987469919025898, + "learning_rate": 1.5789473684210526e-05, + "loss": 0.0001, + "step": 98 + }, + { + "epoch": 0.06943713834823777, + "grad_norm": 0.00033108692150563, + "learning_rate": 1.5964912280701755e-05, + "loss": 0.0, + "step": 99 + }, + { + "epoch": 0.07013852358407856, + "grad_norm": 0.21173974871635437, + "learning_rate": 1.6140350877192984e-05, + "loss": 0.0003, + "step": 100 + }, + { + "epoch": 0.07083990881991933, + "grad_norm": 0.0003154289734084159, + "learning_rate": 1.6315789473684213e-05, + "loss": 0.0, + "step": 101 + }, + { + "epoch": 0.07154129405576012, + "grad_norm": 0.0001488685520598665, + "learning_rate": 1.6491228070175442e-05, + "loss": 0.0, + "step": 102 + }, + { + "epoch": 0.07224267929160091, + "grad_norm": 1.3430032595351804e-05, + "learning_rate": 1.6666666666666667e-05, + "loss": 0.0, + "step": 103 + }, + { + "epoch": 0.07294406452744169, + "grad_norm": 0.011110284365713596, + "learning_rate": 1.6842105263157896e-05, + "loss": 0.0, + "step": 104 + }, + { + "epoch": 0.07364544976328248, + "grad_norm": 1.7930891772266477e-05, + "learning_rate": 1.7017543859649125e-05, + "loss": 0.0, + "step": 105 + }, + { + "epoch": 0.07434683499912327, + "grad_norm": 5.653702146446449e-07, + "learning_rate": 1.719298245614035e-05, + "loss": 0.0, + "step": 106 + }, + { + "epoch": 0.07504822023496406, + "grad_norm": 135.4245147705078, + "learning_rate": 1.736842105263158e-05, + "loss": 6.4792, + "step": 107 + }, + { + "epoch": 0.07574960547080484, + "grad_norm": 153.99264526367188, + "learning_rate": 1.7543859649122806e-05, + "loss": 6.485, + "step": 108 + }, + { + "epoch": 0.07645099070664563, + "grad_norm": 159.35337829589844, + "learning_rate": 1.7719298245614035e-05, + "loss": 2.4936, + "step": 109 + }, + { + "epoch": 0.07715237594248642, + "grad_norm": 0.00017455461784265935, + "learning_rate": 1.7894736842105264e-05, + "loss": 0.0, + "step": 110 + }, + { + "epoch": 0.0778537611783272, + "grad_norm": 146.1802978515625, + "learning_rate": 1.8070175438596493e-05, + "loss": 4.3088, + "step": 111 + }, + { + "epoch": 0.07855514641416798, + "grad_norm": 0.022132212296128273, + "learning_rate": 1.8245614035087722e-05, + "loss": 0.0, + "step": 112 + }, + { + "epoch": 0.07925653165000877, + "grad_norm": 160.43975830078125, + "learning_rate": 1.8421052631578947e-05, + "loss": 4.5829, + "step": 113 + }, + { + "epoch": 0.07995791688584955, + "grad_norm": 164.45138549804688, + "learning_rate": 1.8596491228070176e-05, + "loss": 4.1681, + "step": 114 + }, + { + "epoch": 0.08065930212169034, + "grad_norm": 0.02766432799398899, + "learning_rate": 1.8771929824561405e-05, + "loss": 0.0001, + "step": 115 + }, + { + "epoch": 0.08136068735753113, + "grad_norm": 163.59400939941406, + "learning_rate": 1.8947368421052634e-05, + "loss": 3.0252, + "step": 116 + }, + { + "epoch": 0.08206207259337191, + "grad_norm": 7.11077356338501, + "learning_rate": 1.9122807017543863e-05, + "loss": 0.0309, + "step": 117 + }, + { + "epoch": 0.0827634578292127, + "grad_norm": 15.318146705627441, + "learning_rate": 1.929824561403509e-05, + "loss": 0.0856, + "step": 118 + }, + { + "epoch": 0.08346484306505349, + "grad_norm": 9.1824951171875, + "learning_rate": 1.9473684210526315e-05, + "loss": 0.0416, + "step": 119 + }, + { + "epoch": 0.08416622830089426, + "grad_norm": 22.720653533935547, + "learning_rate": 1.9649122807017544e-05, + "loss": 0.1365, + "step": 120 + }, + { + "epoch": 0.08486761353673505, + "grad_norm": 26.200096130371094, + "learning_rate": 1.9824561403508773e-05, + "loss": 0.2579, + "step": 121 + }, + { + "epoch": 0.08556899877257584, + "grad_norm": 15.874689102172852, + "learning_rate": 2e-05, + "loss": 0.0832, + "step": 122 + }, + { + "epoch": 0.08627038400841662, + "grad_norm": 11.818103790283203, + "learning_rate": 2.0175438596491227e-05, + "loss": 0.0541, + "step": 123 + }, + { + "epoch": 0.08697176924425741, + "grad_norm": 116.5608139038086, + "learning_rate": 2.0350877192982456e-05, + "loss": 0.9623, + "step": 124 + }, + { + "epoch": 0.0876731544800982, + "grad_norm": 10.346467018127441, + "learning_rate": 2.0526315789473685e-05, + "loss": 0.0562, + "step": 125 + }, + { + "epoch": 0.08837453971593898, + "grad_norm": 137.41136169433594, + "learning_rate": 2.0701754385964914e-05, + "loss": 1.1566, + "step": 126 + }, + { + "epoch": 0.08907592495177977, + "grad_norm": 2.9712822437286377, + "learning_rate": 2.0877192982456143e-05, + "loss": 0.0104, + "step": 127 + }, + { + "epoch": 0.08977731018762056, + "grad_norm": 18.723491668701172, + "learning_rate": 2.105263157894737e-05, + "loss": 0.0667, + "step": 128 + }, + { + "epoch": 0.09047869542346133, + "grad_norm": 156.25631713867188, + "learning_rate": 2.1228070175438598e-05, + "loss": 2.4023, + "step": 129 + }, + { + "epoch": 0.09118008065930212, + "grad_norm": 298.97552490234375, + "learning_rate": 2.1403508771929827e-05, + "loss": 6.4928, + "step": 130 + }, + { + "epoch": 0.09188146589514291, + "grad_norm": 87.54296112060547, + "learning_rate": 2.1578947368421053e-05, + "loss": 0.5526, + "step": 131 + }, + { + "epoch": 0.09258285113098369, + "grad_norm": 0.25955039262771606, + "learning_rate": 2.175438596491228e-05, + "loss": 0.0007, + "step": 132 + }, + { + "epoch": 0.09328423636682448, + "grad_norm": 123.57398986816406, + "learning_rate": 2.1929824561403507e-05, + "loss": 1.0402, + "step": 133 + }, + { + "epoch": 0.09398562160266527, + "grad_norm": 238.33229064941406, + "learning_rate": 2.2105263157894736e-05, + "loss": 2.4997, + "step": 134 + }, + { + "epoch": 0.09468700683850605, + "grad_norm": 13.356133460998535, + "learning_rate": 2.2280701754385965e-05, + "loss": 0.0549, + "step": 135 + }, + { + "epoch": 0.09538839207434684, + "grad_norm": 44.56560134887695, + "learning_rate": 2.2456140350877194e-05, + "loss": 0.4078, + "step": 136 + }, + { + "epoch": 0.09608977731018763, + "grad_norm": 25.394866943359375, + "learning_rate": 2.2631578947368423e-05, + "loss": 0.1681, + "step": 137 + }, + { + "epoch": 0.0967911625460284, + "grad_norm": 28.068824768066406, + "learning_rate": 2.280701754385965e-05, + "loss": 0.174, + "step": 138 + }, + { + "epoch": 0.0974925477818692, + "grad_norm": 27.76630401611328, + "learning_rate": 2.2982456140350878e-05, + "loss": 0.1515, + "step": 139 + }, + { + "epoch": 0.09819393301770998, + "grad_norm": 21.082311630249023, + "learning_rate": 2.3157894736842107e-05, + "loss": 0.1379, + "step": 140 + }, + { + "epoch": 0.09889531825355076, + "grad_norm": 16.697647094726562, + "learning_rate": 2.3333333333333336e-05, + "loss": 0.1066, + "step": 141 + }, + { + "epoch": 0.09959670348939155, + "grad_norm": 127.46953582763672, + "learning_rate": 2.3508771929824565e-05, + "loss": 1.3053, + "step": 142 + }, + { + "epoch": 0.10029808872523234, + "grad_norm": 6.2219648361206055, + "learning_rate": 2.368421052631579e-05, + "loss": 0.0323, + "step": 143 + }, + { + "epoch": 0.10099947396107312, + "grad_norm": 0.1309046596288681, + "learning_rate": 2.385964912280702e-05, + "loss": 0.0004, + "step": 144 + }, + { + "epoch": 0.1017008591969139, + "grad_norm": 149.50645446777344, + "learning_rate": 2.4035087719298245e-05, + "loss": 1.8353, + "step": 145 + }, + { + "epoch": 0.1024022444327547, + "grad_norm": 159.4170684814453, + "learning_rate": 2.4210526315789474e-05, + "loss": 3.2714, + "step": 146 + }, + { + "epoch": 0.10310362966859547, + "grad_norm": 0.038805797696113586, + "learning_rate": 2.4385964912280703e-05, + "loss": 0.0001, + "step": 147 + }, + { + "epoch": 0.10380501490443626, + "grad_norm": 301.7401123046875, + "learning_rate": 2.456140350877193e-05, + "loss": 5.5114, + "step": 148 + }, + { + "epoch": 0.10450640014027705, + "grad_norm": 160.01580810546875, + "learning_rate": 2.4736842105263158e-05, + "loss": 3.0398, + "step": 149 + }, + { + "epoch": 0.10520778537611783, + "grad_norm": 1.0593466758728027, + "learning_rate": 2.4912280701754387e-05, + "loss": 0.0026, + "step": 150 + }, + { + "epoch": 0.10590917061195862, + "grad_norm": 0.1588684320449829, + "learning_rate": 2.5087719298245616e-05, + "loss": 0.0004, + "step": 151 + }, + { + "epoch": 0.10661055584779941, + "grad_norm": 0.12831349670886993, + "learning_rate": 2.5263157894736845e-05, + "loss": 0.0004, + "step": 152 + }, + { + "epoch": 0.10731194108364019, + "grad_norm": 1.908389925956726, + "learning_rate": 2.5438596491228074e-05, + "loss": 0.0062, + "step": 153 + }, + { + "epoch": 0.10801332631948098, + "grad_norm": 0.9305810928344727, + "learning_rate": 2.5614035087719303e-05, + "loss": 0.0029, + "step": 154 + }, + { + "epoch": 0.10871471155532177, + "grad_norm": 149.55746459960938, + "learning_rate": 2.578947368421053e-05, + "loss": 2.1251, + "step": 155 + }, + { + "epoch": 0.10941609679116254, + "grad_norm": 83.3541030883789, + "learning_rate": 2.5964912280701754e-05, + "loss": 0.4841, + "step": 156 + }, + { + "epoch": 0.11011748202700333, + "grad_norm": 0.37118402123451233, + "learning_rate": 2.6140350877192983e-05, + "loss": 0.0009, + "step": 157 + }, + { + "epoch": 0.11081886726284412, + "grad_norm": 212.9443359375, + "learning_rate": 2.6315789473684212e-05, + "loss": 1.6639, + "step": 158 + }, + { + "epoch": 0.1115202524986849, + "grad_norm": 1.386982798576355, + "learning_rate": 2.6491228070175438e-05, + "loss": 0.0049, + "step": 159 + }, + { + "epoch": 0.11222163773452569, + "grad_norm": 151.23684692382812, + "learning_rate": 2.6666666666666667e-05, + "loss": 1.913, + "step": 160 + }, + { + "epoch": 0.11292302297036648, + "grad_norm": 276.9901123046875, + "learning_rate": 2.6842105263157896e-05, + "loss": 5.0838, + "step": 161 + }, + { + "epoch": 0.11362440820620726, + "grad_norm": 7.047855377197266, + "learning_rate": 2.7017543859649125e-05, + "loss": 0.0422, + "step": 162 + }, + { + "epoch": 0.11432579344204805, + "grad_norm": 5.565479755401611, + "learning_rate": 2.7192982456140354e-05, + "loss": 0.0252, + "step": 163 + }, + { + "epoch": 0.11502717867788884, + "grad_norm": 11.038246154785156, + "learning_rate": 2.7368421052631583e-05, + "loss": 0.064, + "step": 164 + }, + { + "epoch": 0.11572856391372961, + "grad_norm": 1.4180843830108643, + "learning_rate": 2.754385964912281e-05, + "loss": 0.0048, + "step": 165 + }, + { + "epoch": 0.1164299491495704, + "grad_norm": 131.0841064453125, + "learning_rate": 2.7719298245614034e-05, + "loss": 1.1797, + "step": 166 + }, + { + "epoch": 0.11713133438541119, + "grad_norm": 286.4391784667969, + "learning_rate": 2.7894736842105263e-05, + "loss": 4.2848, + "step": 167 + }, + { + "epoch": 0.11783271962125197, + "grad_norm": 0.354282945394516, + "learning_rate": 2.8070175438596492e-05, + "loss": 0.001, + "step": 168 + }, + { + "epoch": 0.11853410485709276, + "grad_norm": 148.43812561035156, + "learning_rate": 2.824561403508772e-05, + "loss": 2.3637, + "step": 169 + }, + { + "epoch": 0.11923549009293355, + "grad_norm": 3.4153621196746826, + "learning_rate": 2.842105263157895e-05, + "loss": 0.0121, + "step": 170 + }, + { + "epoch": 0.11993687532877433, + "grad_norm": 6.812062740325928, + "learning_rate": 2.8596491228070175e-05, + "loss": 0.038, + "step": 171 + }, + { + "epoch": 0.12063826056461512, + "grad_norm": 0.6381294131278992, + "learning_rate": 2.8771929824561404e-05, + "loss": 0.002, + "step": 172 + }, + { + "epoch": 0.1213396458004559, + "grad_norm": 0.05846899002790451, + "learning_rate": 2.8947368421052634e-05, + "loss": 0.0002, + "step": 173 + }, + { + "epoch": 0.12204103103629668, + "grad_norm": 140.1736297607422, + "learning_rate": 2.9122807017543863e-05, + "loss": 4.0868, + "step": 174 + }, + { + "epoch": 0.12274241627213747, + "grad_norm": 1.7416075468063354, + "learning_rate": 2.929824561403509e-05, + "loss": 0.0057, + "step": 175 + }, + { + "epoch": 0.12344380150797826, + "grad_norm": 0.35068169236183167, + "learning_rate": 2.9473684210526314e-05, + "loss": 0.0009, + "step": 176 + }, + { + "epoch": 0.12414518674381904, + "grad_norm": 135.71665954589844, + "learning_rate": 2.9649122807017543e-05, + "loss": 2.1025, + "step": 177 + }, + { + "epoch": 0.12484657197965983, + "grad_norm": 134.73190307617188, + "learning_rate": 2.9824561403508772e-05, + "loss": 1.8685, + "step": 178 + }, + { + "epoch": 0.1255479572155006, + "grad_norm": 266.5934143066406, + "learning_rate": 3e-05, + "loss": 5.1122, + "step": 179 + }, + { + "epoch": 0.1262493424513414, + "grad_norm": 383.29931640625, + "learning_rate": 3.017543859649123e-05, + "loss": 10.0944, + "step": 180 + }, + { + "epoch": 0.12695072768718219, + "grad_norm": 3.7992992401123047, + "learning_rate": 3.035087719298246e-05, + "loss": 0.0161, + "step": 181 + }, + { + "epoch": 0.12765211292302298, + "grad_norm": 206.0655975341797, + "learning_rate": 3.0526315789473684e-05, + "loss": 1.5303, + "step": 182 + }, + { + "epoch": 0.12835349815886377, + "grad_norm": 13.431243896484375, + "learning_rate": 3.0701754385964913e-05, + "loss": 0.0912, + "step": 183 + }, + { + "epoch": 0.12905488339470453, + "grad_norm": 16.69830894470215, + "learning_rate": 3.087719298245614e-05, + "loss": 0.0824, + "step": 184 + }, + { + "epoch": 0.12975626863054532, + "grad_norm": 13.488637924194336, + "learning_rate": 3.105263157894737e-05, + "loss": 0.0746, + "step": 185 + }, + { + "epoch": 0.1304576538663861, + "grad_norm": 10.065983772277832, + "learning_rate": 3.12280701754386e-05, + "loss": 0.0503, + "step": 186 + }, + { + "epoch": 0.1311590391022269, + "grad_norm": 123.50023651123047, + "learning_rate": 3.140350877192982e-05, + "loss": 1.2874, + "step": 187 + }, + { + "epoch": 0.1318604243380677, + "grad_norm": 1.7797540426254272, + "learning_rate": 3.157894736842105e-05, + "loss": 0.0067, + "step": 188 + }, + { + "epoch": 0.13256180957390848, + "grad_norm": 0.33675551414489746, + "learning_rate": 3.175438596491228e-05, + "loss": 0.0011, + "step": 189 + }, + { + "epoch": 0.13326319480974924, + "grad_norm": 118.4156723022461, + "learning_rate": 3.192982456140351e-05, + "loss": 1.2059, + "step": 190 + }, + { + "epoch": 0.13396458004559003, + "grad_norm": 136.65228271484375, + "learning_rate": 3.210526315789474e-05, + "loss": 1.7856, + "step": 191 + }, + { + "epoch": 0.13466596528143082, + "grad_norm": 145.7415313720703, + "learning_rate": 3.228070175438597e-05, + "loss": 2.8991, + "step": 192 + }, + { + "epoch": 0.1353673505172716, + "grad_norm": 262.7060241699219, + "learning_rate": 3.24561403508772e-05, + "loss": 3.4817, + "step": 193 + }, + { + "epoch": 0.1360687357531124, + "grad_norm": 0.25727707147598267, + "learning_rate": 3.2631578947368426e-05, + "loss": 0.0008, + "step": 194 + }, + { + "epoch": 0.1367701209889532, + "grad_norm": 1.0702918767929077, + "learning_rate": 3.2807017543859655e-05, + "loss": 0.0037, + "step": 195 + }, + { + "epoch": 0.13747150622479395, + "grad_norm": 1.5917607545852661, + "learning_rate": 3.2982456140350884e-05, + "loss": 0.0055, + "step": 196 + }, + { + "epoch": 0.13817289146063474, + "grad_norm": 355.18511962890625, + "learning_rate": 3.3157894736842106e-05, + "loss": 11.4677, + "step": 197 + }, + { + "epoch": 0.13887427669647553, + "grad_norm": 1.7947548627853394, + "learning_rate": 3.3333333333333335e-05, + "loss": 0.0074, + "step": 198 + }, + { + "epoch": 0.13957566193231633, + "grad_norm": 1.5046889781951904, + "learning_rate": 3.3508771929824564e-05, + "loss": 0.0058, + "step": 199 + }, + { + "epoch": 0.14027704716815712, + "grad_norm": 0.2553917467594147, + "learning_rate": 3.368421052631579e-05, + "loss": 0.0008, + "step": 200 + }, + { + "epoch": 0.1409784324039979, + "grad_norm": 0.773563027381897, + "learning_rate": 3.385964912280702e-05, + "loss": 0.0027, + "step": 201 + }, + { + "epoch": 0.14167981763983867, + "grad_norm": 166.71820068359375, + "learning_rate": 3.403508771929825e-05, + "loss": 6.1496, + "step": 202 + }, + { + "epoch": 0.14238120287567946, + "grad_norm": 0.017217637971043587, + "learning_rate": 3.421052631578947e-05, + "loss": 0.0001, + "step": 203 + }, + { + "epoch": 0.14308258811152025, + "grad_norm": 0.0036761912051588297, + "learning_rate": 3.43859649122807e-05, + "loss": 0.0, + "step": 204 + }, + { + "epoch": 0.14378397334736104, + "grad_norm": 0.000679672637488693, + "learning_rate": 3.456140350877193e-05, + "loss": 0.0, + "step": 205 + }, + { + "epoch": 0.14448535858320183, + "grad_norm": 0.00011312704009469599, + "learning_rate": 3.473684210526316e-05, + "loss": 0.0, + "step": 206 + }, + { + "epoch": 0.14518674381904262, + "grad_norm": 0.00012077714200131595, + "learning_rate": 3.491228070175438e-05, + "loss": 0.0, + "step": 207 + }, + { + "epoch": 0.14588812905488338, + "grad_norm": 0.00012628763215616345, + "learning_rate": 3.508771929824561e-05, + "loss": 0.0, + "step": 208 + }, + { + "epoch": 0.14658951429072417, + "grad_norm": 1.7496255168225616e-05, + "learning_rate": 3.526315789473684e-05, + "loss": 0.0, + "step": 209 + }, + { + "epoch": 0.14729089952656496, + "grad_norm": 1.5739196896902286e-05, + "learning_rate": 3.543859649122807e-05, + "loss": 0.0, + "step": 210 + }, + { + "epoch": 0.14799228476240575, + "grad_norm": 0.08704076707363129, + "learning_rate": 3.56140350877193e-05, + "loss": 0.0001, + "step": 211 + }, + { + "epoch": 0.14869366999824654, + "grad_norm": 0.0016435593133792281, + "learning_rate": 3.578947368421053e-05, + "loss": 0.0, + "step": 212 + }, + { + "epoch": 0.14939505523408733, + "grad_norm": 1.5830031633377075, + "learning_rate": 3.5964912280701756e-05, + "loss": 0.005, + "step": 213 + }, + { + "epoch": 0.15009644046992812, + "grad_norm": 0.00016859408060554415, + "learning_rate": 3.6140350877192985e-05, + "loss": 0.0, + "step": 214 + }, + { + "epoch": 0.15079782570576888, + "grad_norm": 1.2645250535570085e-06, + "learning_rate": 3.6315789473684214e-05, + "loss": 0.0, + "step": 215 + }, + { + "epoch": 0.15149921094160967, + "grad_norm": 336.5249328613281, + "learning_rate": 3.6491228070175443e-05, + "loss": 4.1939, + "step": 216 + }, + { + "epoch": 0.15220059617745046, + "grad_norm": 9.010884127746976e-07, + "learning_rate": 3.6666666666666666e-05, + "loss": 0.0, + "step": 217 + }, + { + "epoch": 0.15290198141329125, + "grad_norm": 5.530741304937692e-07, + "learning_rate": 3.6842105263157895e-05, + "loss": 0.0, + "step": 218 + }, + { + "epoch": 0.15360336664913204, + "grad_norm": 222.7361297607422, + "learning_rate": 3.7017543859649124e-05, + "loss": 0.7712, + "step": 219 + }, + { + "epoch": 0.15430475188497284, + "grad_norm": 123.51704406738281, + "learning_rate": 3.719298245614035e-05, + "loss": 7.8703, + "step": 220 + }, + { + "epoch": 0.1550061371208136, + "grad_norm": 119.08999633789062, + "learning_rate": 3.736842105263158e-05, + "loss": 6.6427, + "step": 221 + }, + { + "epoch": 0.1557075223566544, + "grad_norm": 0.005513960495591164, + "learning_rate": 3.754385964912281e-05, + "loss": 0.0, + "step": 222 + }, + { + "epoch": 0.15640890759249518, + "grad_norm": 0.021047895774245262, + "learning_rate": 3.771929824561404e-05, + "loss": 0.0, + "step": 223 + }, + { + "epoch": 0.15711029282833597, + "grad_norm": 0.001549599110148847, + "learning_rate": 3.789473684210527e-05, + "loss": 0.0, + "step": 224 + }, + { + "epoch": 0.15781167806417676, + "grad_norm": 0.008736948482692242, + "learning_rate": 3.80701754385965e-05, + "loss": 0.0, + "step": 225 + }, + { + "epoch": 0.15851306330001755, + "grad_norm": 0.03641926124691963, + "learning_rate": 3.824561403508773e-05, + "loss": 0.0001, + "step": 226 + }, + { + "epoch": 0.1592144485358583, + "grad_norm": 0.026789208874106407, + "learning_rate": 3.842105263157895e-05, + "loss": 0.0001, + "step": 227 + }, + { + "epoch": 0.1599158337716991, + "grad_norm": 130.79087829589844, + "learning_rate": 3.859649122807018e-05, + "loss": 3.4204, + "step": 228 + }, + { + "epoch": 0.1606172190075399, + "grad_norm": 404.3518981933594, + "learning_rate": 3.877192982456141e-05, + "loss": 5.6741, + "step": 229 + }, + { + "epoch": 0.16131860424338068, + "grad_norm": 5.39788818359375, + "learning_rate": 3.894736842105263e-05, + "loss": 0.0251, + "step": 230 + }, + { + "epoch": 0.16201998947922147, + "grad_norm": 18.25702667236328, + "learning_rate": 3.912280701754386e-05, + "loss": 0.1059, + "step": 231 + }, + { + "epoch": 0.16272137471506226, + "grad_norm": 2.5649735927581787, + "learning_rate": 3.929824561403509e-05, + "loss": 0.0108, + "step": 232 + }, + { + "epoch": 0.16342275995090302, + "grad_norm": 389.410400390625, + "learning_rate": 3.9473684210526316e-05, + "loss": 4.1361, + "step": 233 + }, + { + "epoch": 0.16412414518674381, + "grad_norm": 97.98786926269531, + "learning_rate": 3.9649122807017545e-05, + "loss": 1.0207, + "step": 234 + }, + { + "epoch": 0.1648255304225846, + "grad_norm": 446.8147277832031, + "learning_rate": 3.9824561403508774e-05, + "loss": 2.3678, + "step": 235 + }, + { + "epoch": 0.1655269156584254, + "grad_norm": 88.88761901855469, + "learning_rate": 4e-05, + "loss": 0.814, + "step": 236 + }, + { + "epoch": 0.16622830089426618, + "grad_norm": 217.3586883544922, + "learning_rate": 4.017543859649123e-05, + "loss": 9.7291, + "step": 237 + }, + { + "epoch": 0.16692968613010697, + "grad_norm": 106.2865219116211, + "learning_rate": 4.0350877192982455e-05, + "loss": 1.1716, + "step": 238 + }, + { + "epoch": 0.16763107136594774, + "grad_norm": 154.2283477783203, + "learning_rate": 4.0526315789473684e-05, + "loss": 0.6351, + "step": 239 + }, + { + "epoch": 0.16833245660178853, + "grad_norm": 37.964447021484375, + "learning_rate": 4.070175438596491e-05, + "loss": 0.3484, + "step": 240 + }, + { + "epoch": 0.16903384183762932, + "grad_norm": 24.696714401245117, + "learning_rate": 4.087719298245614e-05, + "loss": 0.2762, + "step": 241 + }, + { + "epoch": 0.1697352270734701, + "grad_norm": 82.12918853759766, + "learning_rate": 4.105263157894737e-05, + "loss": 0.9978, + "step": 242 + }, + { + "epoch": 0.1704366123093109, + "grad_norm": 41.84835433959961, + "learning_rate": 4.12280701754386e-05, + "loss": 0.4799, + "step": 243 + }, + { + "epoch": 0.1711379975451517, + "grad_norm": 45.34370803833008, + "learning_rate": 4.140350877192983e-05, + "loss": 0.4255, + "step": 244 + }, + { + "epoch": 0.17183938278099245, + "grad_norm": 41.85076904296875, + "learning_rate": 4.157894736842106e-05, + "loss": 0.616, + "step": 245 + }, + { + "epoch": 0.17254076801683324, + "grad_norm": 31.718475341796875, + "learning_rate": 4.1754385964912287e-05, + "loss": 0.2034, + "step": 246 + }, + { + "epoch": 0.17324215325267403, + "grad_norm": 149.61940002441406, + "learning_rate": 4.1929824561403516e-05, + "loss": 1.5976, + "step": 247 + }, + { + "epoch": 0.17394353848851482, + "grad_norm": 127.43466186523438, + "learning_rate": 4.210526315789474e-05, + "loss": 1.4465, + "step": 248 + }, + { + "epoch": 0.1746449237243556, + "grad_norm": 36.885379791259766, + "learning_rate": 4.228070175438597e-05, + "loss": 0.5706, + "step": 249 + }, + { + "epoch": 0.1753463089601964, + "grad_norm": 343.46649169921875, + "learning_rate": 4.2456140350877196e-05, + "loss": 5.0778, + "step": 250 + }, + { + "epoch": 0.17604769419603716, + "grad_norm": 23.604751586914062, + "learning_rate": 4.2631578947368425e-05, + "loss": 0.1635, + "step": 251 + }, + { + "epoch": 0.17674907943187795, + "grad_norm": 37.630043029785156, + "learning_rate": 4.2807017543859654e-05, + "loss": 0.5644, + "step": 252 + }, + { + "epoch": 0.17745046466771874, + "grad_norm": 41.07538986206055, + "learning_rate": 4.298245614035088e-05, + "loss": 0.5337, + "step": 253 + }, + { + "epoch": 0.17815184990355953, + "grad_norm": 66.12769317626953, + "learning_rate": 4.3157894736842105e-05, + "loss": 0.8491, + "step": 254 + }, + { + "epoch": 0.17885323513940032, + "grad_norm": 38.53345489501953, + "learning_rate": 4.3333333333333334e-05, + "loss": 0.4, + "step": 255 + }, + { + "epoch": 0.17955462037524111, + "grad_norm": 37.57467269897461, + "learning_rate": 4.350877192982456e-05, + "loss": 0.6272, + "step": 256 + }, + { + "epoch": 0.18025600561108188, + "grad_norm": 27.241796493530273, + "learning_rate": 4.368421052631579e-05, + "loss": 0.2307, + "step": 257 + }, + { + "epoch": 0.18095739084692267, + "grad_norm": 24.353778839111328, + "learning_rate": 4.3859649122807014e-05, + "loss": 0.1238, + "step": 258 + }, + { + "epoch": 0.18165877608276346, + "grad_norm": 5.489215850830078, + "learning_rate": 4.403508771929824e-05, + "loss": 0.031, + "step": 259 + }, + { + "epoch": 0.18236016131860425, + "grad_norm": 16.305925369262695, + "learning_rate": 4.421052631578947e-05, + "loss": 0.0398, + "step": 260 + }, + { + "epoch": 0.18306154655444504, + "grad_norm": 6.403785228729248, + "learning_rate": 4.43859649122807e-05, + "loss": 0.0304, + "step": 261 + }, + { + "epoch": 0.18376293179028583, + "grad_norm": 4.471805572509766, + "learning_rate": 4.456140350877193e-05, + "loss": 0.0077, + "step": 262 + }, + { + "epoch": 0.1844643170261266, + "grad_norm": 0.9698334336280823, + "learning_rate": 4.473684210526316e-05, + "loss": 0.0031, + "step": 263 + }, + { + "epoch": 0.18516570226196738, + "grad_norm": 0.014244407415390015, + "learning_rate": 4.491228070175439e-05, + "loss": 0.0, + "step": 264 + }, + { + "epoch": 0.18586708749780817, + "grad_norm": 121.71415710449219, + "learning_rate": 4.508771929824562e-05, + "loss": 4.2289, + "step": 265 + }, + { + "epoch": 0.18656847273364896, + "grad_norm": 0.02348833531141281, + "learning_rate": 4.5263157894736846e-05, + "loss": 0.0001, + "step": 266 + }, + { + "epoch": 0.18726985796948975, + "grad_norm": 0.0016702099237591028, + "learning_rate": 4.5438596491228075e-05, + "loss": 0.0, + "step": 267 + }, + { + "epoch": 0.18797124320533054, + "grad_norm": 0.0027486609760671854, + "learning_rate": 4.56140350877193e-05, + "loss": 0.0, + "step": 268 + }, + { + "epoch": 0.1886726284411713, + "grad_norm": 0.0012574723223224282, + "learning_rate": 4.5789473684210527e-05, + "loss": 0.0, + "step": 269 + }, + { + "epoch": 0.1893740136770121, + "grad_norm": 0.002944325562566519, + "learning_rate": 4.5964912280701756e-05, + "loss": 0.0, + "step": 270 + }, + { + "epoch": 0.19007539891285288, + "grad_norm": 0.0020754581782966852, + "learning_rate": 4.6140350877192985e-05, + "loss": 0.0, + "step": 271 + }, + { + "epoch": 0.19077678414869367, + "grad_norm": 0.008111892268061638, + "learning_rate": 4.6315789473684214e-05, + "loss": 0.0, + "step": 272 + }, + { + "epoch": 0.19147816938453446, + "grad_norm": 118.25870513916016, + "learning_rate": 4.649122807017544e-05, + "loss": 3.5798, + "step": 273 + }, + { + "epoch": 0.19217955462037525, + "grad_norm": 118.91422271728516, + "learning_rate": 4.666666666666667e-05, + "loss": 2.0933, + "step": 274 + }, + { + "epoch": 0.19288093985621602, + "grad_norm": 0.1281914860010147, + "learning_rate": 4.68421052631579e-05, + "loss": 0.0004, + "step": 275 + }, + { + "epoch": 0.1935823250920568, + "grad_norm": 0.9166204929351807, + "learning_rate": 4.701754385964913e-05, + "loss": 0.0034, + "step": 276 + }, + { + "epoch": 0.1942837103278976, + "grad_norm": 3.7037837505340576, + "learning_rate": 4.719298245614036e-05, + "loss": 0.0179, + "step": 277 + }, + { + "epoch": 0.1949850955637384, + "grad_norm": 0.716585099697113, + "learning_rate": 4.736842105263158e-05, + "loss": 0.0028, + "step": 278 + }, + { + "epoch": 0.19568648079957918, + "grad_norm": 119.46430969238281, + "learning_rate": 4.754385964912281e-05, + "loss": 1.7075, + "step": 279 + }, + { + "epoch": 0.19638786603541997, + "grad_norm": 0.07794260233640671, + "learning_rate": 4.771929824561404e-05, + "loss": 0.0003, + "step": 280 + }, + { + "epoch": 0.19708925127126073, + "grad_norm": 119.65841674804688, + "learning_rate": 4.789473684210526e-05, + "loss": 3.3725, + "step": 281 + }, + { + "epoch": 0.19779063650710152, + "grad_norm": 0.06007947400212288, + "learning_rate": 4.807017543859649e-05, + "loss": 0.0002, + "step": 282 + }, + { + "epoch": 0.1984920217429423, + "grad_norm": 0.33029523491859436, + "learning_rate": 4.824561403508772e-05, + "loss": 0.0013, + "step": 283 + }, + { + "epoch": 0.1991934069787831, + "grad_norm": 0.19337859749794006, + "learning_rate": 4.842105263157895e-05, + "loss": 0.0007, + "step": 284 + }, + { + "epoch": 0.1998947922146239, + "grad_norm": 0.4252503514289856, + "learning_rate": 4.859649122807018e-05, + "loss": 0.0017, + "step": 285 + }, + { + "epoch": 0.20059617745046468, + "grad_norm": 0.07317744940519333, + "learning_rate": 4.8771929824561406e-05, + "loss": 0.0003, + "step": 286 + }, + { + "epoch": 0.20129756268630544, + "grad_norm": 0.01806594245135784, + "learning_rate": 4.8947368421052635e-05, + "loss": 0.0001, + "step": 287 + }, + { + "epoch": 0.20199894792214623, + "grad_norm": 106.39412689208984, + "learning_rate": 4.912280701754386e-05, + "loss": 3.9844, + "step": 288 + }, + { + "epoch": 0.20270033315798702, + "grad_norm": 0.008989217691123486, + "learning_rate": 4.9298245614035086e-05, + "loss": 0.0, + "step": 289 + }, + { + "epoch": 0.2034017183938278, + "grad_norm": 0.05597059056162834, + "learning_rate": 4.9473684210526315e-05, + "loss": 0.0002, + "step": 290 + }, + { + "epoch": 0.2041031036296686, + "grad_norm": 113.05389404296875, + "learning_rate": 4.9649122807017544e-05, + "loss": 2.5036, + "step": 291 + }, + { + "epoch": 0.2048044888655094, + "grad_norm": 0.39913010597229004, + "learning_rate": 4.9824561403508773e-05, + "loss": 0.0015, + "step": 292 + }, + { + "epoch": 0.20550587410135016, + "grad_norm": 0.944106936454773, + "learning_rate": 5e-05, + "loss": 0.0038, + "step": 293 + }, + { + "epoch": 0.20620725933719095, + "grad_norm": 108.76079559326172, + "learning_rate": 4.9999995795715716e-05, + "loss": 1.4773, + "step": 294 + }, + { + "epoch": 0.20690864457303174, + "grad_norm": 1.2869099378585815, + "learning_rate": 4.999998318286425e-05, + "loss": 0.0054, + "step": 295 + }, + { + "epoch": 0.20761002980887253, + "grad_norm": 0.7701173424720764, + "learning_rate": 4.999996216144987e-05, + "loss": 0.0029, + "step": 296 + }, + { + "epoch": 0.20831141504471332, + "grad_norm": 104.06690216064453, + "learning_rate": 4.9999932731479625e-05, + "loss": 1.2931, + "step": 297 + }, + { + "epoch": 0.2090128002805541, + "grad_norm": 234.58531188964844, + "learning_rate": 4.999989489296344e-05, + "loss": 4.5688, + "step": 298 + }, + { + "epoch": 0.20971418551639487, + "grad_norm": 108.04949188232422, + "learning_rate": 4.999984864591401e-05, + "loss": 1.5544, + "step": 299 + }, + { + "epoch": 0.21041557075223566, + "grad_norm": 1.8640260696411133, + "learning_rate": 4.999979399034691e-05, + "loss": 0.0078, + "step": 300 + }, + { + "epoch": 0.21111695598807645, + "grad_norm": 3.089526414871216, + "learning_rate": 4.999973092628052e-05, + "loss": 0.0138, + "step": 301 + }, + { + "epoch": 0.21181834122391724, + "grad_norm": 1.4183728694915771, + "learning_rate": 4.999965945373605e-05, + "loss": 0.006, + "step": 302 + }, + { + "epoch": 0.21251972645975803, + "grad_norm": 0.9733534455299377, + "learning_rate": 4.9999579572737533e-05, + "loss": 0.0038, + "step": 303 + }, + { + "epoch": 0.21322111169559882, + "grad_norm": 121.11236572265625, + "learning_rate": 4.9999491283311836e-05, + "loss": 2.2543, + "step": 304 + }, + { + "epoch": 0.21392249693143958, + "grad_norm": 117.718017578125, + "learning_rate": 4.999939458548868e-05, + "loss": 2.5007, + "step": 305 + }, + { + "epoch": 0.21462388216728037, + "grad_norm": 0.013796065002679825, + "learning_rate": 4.9999289479300557e-05, + "loss": 0.0, + "step": 306 + }, + { + "epoch": 0.21532526740312116, + "grad_norm": 116.46581268310547, + "learning_rate": 4.999917596478283e-05, + "loss": 2.6005, + "step": 307 + }, + { + "epoch": 0.21602665263896195, + "grad_norm": 0.3713426887989044, + "learning_rate": 4.999905404197368e-05, + "loss": 0.0014, + "step": 308 + }, + { + "epoch": 0.21672803787480274, + "grad_norm": 0.44403278827667236, + "learning_rate": 4.999892371091411e-05, + "loss": 0.0017, + "step": 309 + }, + { + "epoch": 0.21742942311064353, + "grad_norm": 139.35816955566406, + "learning_rate": 4.999878497164797e-05, + "loss": 11.1347, + "step": 310 + }, + { + "epoch": 0.2181308083464843, + "grad_norm": 0.20512554049491882, + "learning_rate": 4.999863782422191e-05, + "loss": 0.0007, + "step": 311 + }, + { + "epoch": 0.21883219358232509, + "grad_norm": 88.1954345703125, + "learning_rate": 4.9998482268685434e-05, + "loss": 0.843, + "step": 312 + }, + { + "epoch": 0.21953357881816588, + "grad_norm": 93.83118438720703, + "learning_rate": 4.999831830509084e-05, + "loss": 1.0804, + "step": 313 + }, + { + "epoch": 0.22023496405400667, + "grad_norm": 5.042333602905273, + "learning_rate": 4.99981459334933e-05, + "loss": 0.0256, + "step": 314 + }, + { + "epoch": 0.22093634928984746, + "grad_norm": 7.477559566497803, + "learning_rate": 4.9997965153950775e-05, + "loss": 0.0402, + "step": 315 + }, + { + "epoch": 0.22163773452568825, + "grad_norm": 156.78961181640625, + "learning_rate": 4.999777596652408e-05, + "loss": 6.0824, + "step": 316 + }, + { + "epoch": 0.222339119761529, + "grad_norm": 9.518619537353516, + "learning_rate": 4.999757837127683e-05, + "loss": 0.0577, + "step": 317 + }, + { + "epoch": 0.2230405049973698, + "grad_norm": 2.860729455947876, + "learning_rate": 4.9997372368275506e-05, + "loss": 0.0135, + "step": 318 + }, + { + "epoch": 0.2237418902332106, + "grad_norm": 1.1225600242614746, + "learning_rate": 4.999715795758938e-05, + "loss": 0.0049, + "step": 319 + }, + { + "epoch": 0.22444327546905138, + "grad_norm": 0.12374955415725708, + "learning_rate": 4.9996935139290575e-05, + "loss": 0.0004, + "step": 320 + }, + { + "epoch": 0.22514466070489217, + "grad_norm": 112.18521881103516, + "learning_rate": 4.999670391345403e-05, + "loss": 3.4372, + "step": 321 + }, + { + "epoch": 0.22584604594073296, + "grad_norm": 0.0004930093418806791, + "learning_rate": 4.999646428015752e-05, + "loss": 0.0, + "step": 322 + }, + { + "epoch": 0.22654743117657372, + "grad_norm": 0.00027333476464264095, + "learning_rate": 4.9996216239481643e-05, + "loss": 0.0, + "step": 323 + }, + { + "epoch": 0.2272488164124145, + "grad_norm": 0.0001735862751957029, + "learning_rate": 4.999595979150982e-05, + "loss": 0.0, + "step": 324 + }, + { + "epoch": 0.2279502016482553, + "grad_norm": 2.878743180190213e-05, + "learning_rate": 4.9995694936328316e-05, + "loss": 0.0, + "step": 325 + }, + { + "epoch": 0.2286515868840961, + "grad_norm": 109.58158111572266, + "learning_rate": 4.99954216740262e-05, + "loss": 4.991, + "step": 326 + }, + { + "epoch": 0.22935297211993688, + "grad_norm": 113.2159194946289, + "learning_rate": 4.9995140004695396e-05, + "loss": 4.7235, + "step": 327 + }, + { + "epoch": 0.23005435735577767, + "grad_norm": 0.00015405479643959552, + "learning_rate": 4.999484992843063e-05, + "loss": 0.0, + "step": 328 + }, + { + "epoch": 0.23075574259161843, + "grad_norm": 0.011504331603646278, + "learning_rate": 4.999455144532947e-05, + "loss": 0.0, + "step": 329 + }, + { + "epoch": 0.23145712782745922, + "grad_norm": 0.001587590086273849, + "learning_rate": 4.9994244555492315e-05, + "loss": 0.0, + "step": 330 + }, + { + "epoch": 0.23215851306330001, + "grad_norm": 0.0038678112905472517, + "learning_rate": 4.999392925902238e-05, + "loss": 0.0, + "step": 331 + }, + { + "epoch": 0.2328598982991408, + "grad_norm": 0.044445931911468506, + "learning_rate": 4.999360555602571e-05, + "loss": 0.0001, + "step": 332 + }, + { + "epoch": 0.2335612835349816, + "grad_norm": 108.20228576660156, + "learning_rate": 4.999327344661118e-05, + "loss": 2.0004, + "step": 333 + }, + { + "epoch": 0.23426266877082239, + "grad_norm": 0.17047792673110962, + "learning_rate": 4.9992932930890505e-05, + "loss": 0.0005, + "step": 334 + }, + { + "epoch": 0.23496405400666315, + "grad_norm": 0.39230290055274963, + "learning_rate": 4.999258400897819e-05, + "loss": 0.0016, + "step": 335 + }, + { + "epoch": 0.23566543924250394, + "grad_norm": 191.8765106201172, + "learning_rate": 4.999222668099163e-05, + "loss": 2.4042, + "step": 336 + }, + { + "epoch": 0.23636682447834473, + "grad_norm": 87.04983520507812, + "learning_rate": 4.999186094705097e-05, + "loss": 0.9342, + "step": 337 + }, + { + "epoch": 0.23706820971418552, + "grad_norm": 14.748551368713379, + "learning_rate": 4.999148680727925e-05, + "loss": 0.1165, + "step": 338 + }, + { + "epoch": 0.2377695949500263, + "grad_norm": 16.103792190551758, + "learning_rate": 4.999110426180229e-05, + "loss": 0.11, + "step": 339 + }, + { + "epoch": 0.2384709801858671, + "grad_norm": 4.507053375244141, + "learning_rate": 4.999071331074877e-05, + "loss": 0.1421, + "step": 340 + }, + { + "epoch": 0.23917236542170786, + "grad_norm": 8.822351455688477, + "learning_rate": 4.999031395425018e-05, + "loss": 0.2372, + "step": 341 + }, + { + "epoch": 0.23987375065754865, + "grad_norm": 19.92313575744629, + "learning_rate": 4.998990619244084e-05, + "loss": 0.1537, + "step": 342 + }, + { + "epoch": 0.24057513589338944, + "grad_norm": 12.00966739654541, + "learning_rate": 4.998949002545789e-05, + "loss": 0.0872, + "step": 343 + }, + { + "epoch": 0.24127652112923023, + "grad_norm": 4.180269718170166, + "learning_rate": 4.998906545344132e-05, + "loss": 0.0218, + "step": 344 + }, + { + "epoch": 0.24197790636507102, + "grad_norm": 0.596821129322052, + "learning_rate": 4.998863247653392e-05, + "loss": 0.0025, + "step": 345 + }, + { + "epoch": 0.2426792916009118, + "grad_norm": 0.4522903263568878, + "learning_rate": 4.998819109488132e-05, + "loss": 0.0013, + "step": 346 + }, + { + "epoch": 0.24338067683675257, + "grad_norm": 0.0007598726078867912, + "learning_rate": 4.998774130863199e-05, + "loss": 0.0, + "step": 347 + }, + { + "epoch": 0.24408206207259336, + "grad_norm": 8.103512664092705e-05, + "learning_rate": 4.998728311793719e-05, + "loss": 0.0, + "step": 348 + }, + { + "epoch": 0.24478344730843415, + "grad_norm": 105.80138397216797, + "learning_rate": 4.998681652295104e-05, + "loss": 4.8223, + "step": 349 + }, + { + "epoch": 0.24548483254427494, + "grad_norm": 127.92369079589844, + "learning_rate": 4.998634152383047e-05, + "loss": 7.0698, + "step": 350 + }, + { + "epoch": 0.24618621778011573, + "grad_norm": 2.636051704030251e-06, + "learning_rate": 4.998585812073525e-05, + "loss": 0.0, + "step": 351 + }, + { + "epoch": 0.24688760301595652, + "grad_norm": 2.710710077735712e-06, + "learning_rate": 4.9985366313827975e-05, + "loss": 0.0, + "step": 352 + }, + { + "epoch": 0.2475889882517973, + "grad_norm": 103.96717834472656, + "learning_rate": 4.998486610327405e-05, + "loss": 6.485, + "step": 353 + }, + { + "epoch": 0.24829037348763808, + "grad_norm": 103.35551452636719, + "learning_rate": 4.9984357489241715e-05, + "loss": 5.7037, + "step": 354 + }, + { + "epoch": 0.24899175872347887, + "grad_norm": 100.56452178955078, + "learning_rate": 4.998384047190204e-05, + "loss": 4.1908, + "step": 355 + }, + { + "epoch": 0.24969314395931966, + "grad_norm": 200.88751220703125, + "learning_rate": 4.998331505142893e-05, + "loss": 4.8984, + "step": 356 + }, + { + "epoch": 0.2503945291951604, + "grad_norm": 77.69247436523438, + "learning_rate": 4.99827812279991e-05, + "loss": 0.9449, + "step": 357 + }, + { + "epoch": 0.2510959144310012, + "grad_norm": 26.622453689575195, + "learning_rate": 4.9982239001792095e-05, + "loss": 0.3613, + "step": 358 + }, + { + "epoch": 0.251797299666842, + "grad_norm": 26.442935943603516, + "learning_rate": 4.9981688372990284e-05, + "loss": 0.5424, + "step": 359 + }, + { + "epoch": 0.2524986849026828, + "grad_norm": 39.277618408203125, + "learning_rate": 4.9981129341778886e-05, + "loss": 0.9456, + "step": 360 + }, + { + "epoch": 0.2532000701385236, + "grad_norm": 43.96021270751953, + "learning_rate": 4.9980561908345916e-05, + "loss": 0.6765, + "step": 361 + }, + { + "epoch": 0.25390145537436437, + "grad_norm": 14.254291534423828, + "learning_rate": 4.997998607288222e-05, + "loss": 0.1276, + "step": 362 + }, + { + "epoch": 0.25460284061020516, + "grad_norm": 0.6544491052627563, + "learning_rate": 4.9979401835581476e-05, + "loss": 0.0032, + "step": 363 + }, + { + "epoch": 0.25530422584604595, + "grad_norm": 0.011080354452133179, + "learning_rate": 4.99788091966402e-05, + "loss": 0.0, + "step": 364 + }, + { + "epoch": 0.25600561108188674, + "grad_norm": 284.8055114746094, + "learning_rate": 4.997820815625771e-05, + "loss": 7.6058, + "step": 365 + }, + { + "epoch": 0.25670699631772753, + "grad_norm": 89.02140045166016, + "learning_rate": 4.997759871463618e-05, + "loss": 5.079, + "step": 366 + }, + { + "epoch": 0.2574083815535683, + "grad_norm": 4.7222201828844845e-05, + "learning_rate": 4.997698087198056e-05, + "loss": 0.0, + "step": 367 + }, + { + "epoch": 0.25810976678940906, + "grad_norm": 85.66144561767578, + "learning_rate": 4.997635462849869e-05, + "loss": 4.8183, + "step": 368 + }, + { + "epoch": 0.25881115202524985, + "grad_norm": 0.0015366720035672188, + "learning_rate": 4.997571998440118e-05, + "loss": 0.0, + "step": 369 + }, + { + "epoch": 0.25951253726109064, + "grad_norm": 100.14698028564453, + "learning_rate": 4.99750769399015e-05, + "loss": 2.2663, + "step": 370 + }, + { + "epoch": 0.2602139224969314, + "grad_norm": 0.4829590618610382, + "learning_rate": 4.997442549521592e-05, + "loss": 0.0021, + "step": 371 + }, + { + "epoch": 0.2609153077327722, + "grad_norm": 70.44209289550781, + "learning_rate": 4.9973765650563564e-05, + "loss": 0.7532, + "step": 372 + }, + { + "epoch": 0.261616692968613, + "grad_norm": 272.1163024902344, + "learning_rate": 4.9973097406166355e-05, + "loss": 3.9935, + "step": 373 + }, + { + "epoch": 0.2623180782044538, + "grad_norm": 25.12706184387207, + "learning_rate": 4.997242076224906e-05, + "loss": 0.3738, + "step": 374 + }, + { + "epoch": 0.2630194634402946, + "grad_norm": 35.15785598754883, + "learning_rate": 4.997173571903926e-05, + "loss": 0.6808, + "step": 375 + }, + { + "epoch": 0.2637208486761354, + "grad_norm": 27.46406364440918, + "learning_rate": 4.9971042276767366e-05, + "loss": 0.2568, + "step": 376 + }, + { + "epoch": 0.26442223391197617, + "grad_norm": 6.682714462280273, + "learning_rate": 4.997034043566661e-05, + "loss": 0.0374, + "step": 377 + }, + { + "epoch": 0.26512361914781696, + "grad_norm": 0.057287223637104034, + "learning_rate": 4.9969630195973045e-05, + "loss": 0.0001, + "step": 378 + }, + { + "epoch": 0.26582500438365775, + "grad_norm": 0.006871811114251614, + "learning_rate": 4.9968911557925564e-05, + "loss": 0.0, + "step": 379 + }, + { + "epoch": 0.2665263896194985, + "grad_norm": 0.0037054684944450855, + "learning_rate": 4.996818452176587e-05, + "loss": 0.0, + "step": 380 + }, + { + "epoch": 0.2672277748553393, + "grad_norm": 123.15022277832031, + "learning_rate": 4.9967449087738505e-05, + "loss": 5.8346, + "step": 381 + }, + { + "epoch": 0.26792916009118006, + "grad_norm": 121.81306457519531, + "learning_rate": 4.996670525609082e-05, + "loss": 4.3778, + "step": 382 + }, + { + "epoch": 0.26863054532702085, + "grad_norm": 0.00026572938077151775, + "learning_rate": 4.9965953027072996e-05, + "loss": 0.0, + "step": 383 + }, + { + "epoch": 0.26933193056286164, + "grad_norm": 198.09776306152344, + "learning_rate": 4.996519240093806e-05, + "loss": 5.2574, + "step": 384 + }, + { + "epoch": 0.27003331579870243, + "grad_norm": 130.21214294433594, + "learning_rate": 4.996442337794181e-05, + "loss": 3.217, + "step": 385 + }, + { + "epoch": 0.2707347010345432, + "grad_norm": 133.94178771972656, + "learning_rate": 4.996364595834291e-05, + "loss": 1.3412, + "step": 386 + }, + { + "epoch": 0.271436086270384, + "grad_norm": 12.223856925964355, + "learning_rate": 4.9962860142402853e-05, + "loss": 0.0924, + "step": 387 + }, + { + "epoch": 0.2721374715062248, + "grad_norm": 21.94683074951172, + "learning_rate": 4.9962065930385934e-05, + "loss": 0.1824, + "step": 388 + }, + { + "epoch": 0.2728388567420656, + "grad_norm": 46.802406311035156, + "learning_rate": 4.996126332255929e-05, + "loss": 0.3585, + "step": 389 + }, + { + "epoch": 0.2735402419779064, + "grad_norm": 7.21342134475708, + "learning_rate": 4.9960452319192866e-05, + "loss": 0.1045, + "step": 390 + }, + { + "epoch": 0.2742416272137472, + "grad_norm": 33.5169677734375, + "learning_rate": 4.995963292055943e-05, + "loss": 0.248, + "step": 391 + }, + { + "epoch": 0.2749430124495879, + "grad_norm": 16.911136627197266, + "learning_rate": 4.995880512693458e-05, + "loss": 0.0876, + "step": 392 + }, + { + "epoch": 0.2756443976854287, + "grad_norm": 206.7276153564453, + "learning_rate": 4.995796893859675e-05, + "loss": 1.6484, + "step": 393 + }, + { + "epoch": 0.2763457829212695, + "grad_norm": 94.29922485351562, + "learning_rate": 4.9957124355827186e-05, + "loss": 0.8745, + "step": 394 + }, + { + "epoch": 0.2770471681571103, + "grad_norm": 156.6600799560547, + "learning_rate": 4.9956271378909945e-05, + "loss": 0.9049, + "step": 395 + }, + { + "epoch": 0.27774855339295107, + "grad_norm": 6.1268744468688965, + "learning_rate": 4.995541000813193e-05, + "loss": 0.0343, + "step": 396 + }, + { + "epoch": 0.27844993862879186, + "grad_norm": 1.731136679649353, + "learning_rate": 4.995454024378285e-05, + "loss": 0.0075, + "step": 397 + }, + { + "epoch": 0.27915132386463265, + "grad_norm": 144.5174560546875, + "learning_rate": 4.995366208615526e-05, + "loss": 2.6072, + "step": 398 + }, + { + "epoch": 0.27985270910047344, + "grad_norm": 0.19711287319660187, + "learning_rate": 4.9952775535544494e-05, + "loss": 0.0007, + "step": 399 + }, + { + "epoch": 0.28055409433631423, + "grad_norm": 0.13065628707408905, + "learning_rate": 4.9951880592248754e-05, + "loss": 0.0004, + "step": 400 + }, + { + "epoch": 0.281255479572155, + "grad_norm": 125.28437042236328, + "learning_rate": 4.995097725656905e-05, + "loss": 1.8792, + "step": 401 + }, + { + "epoch": 0.2819568648079958, + "grad_norm": 111.38099670410156, + "learning_rate": 4.9950065528809214e-05, + "loss": 2.548, + "step": 402 + }, + { + "epoch": 0.2826582500438366, + "grad_norm": 1.2473199367523193, + "learning_rate": 4.994914540927589e-05, + "loss": 0.0042, + "step": 403 + }, + { + "epoch": 0.28335963527967734, + "grad_norm": 0.7474699020385742, + "learning_rate": 4.994821689827856e-05, + "loss": 0.0026, + "step": 404 + }, + { + "epoch": 0.2840610205155181, + "grad_norm": 0.14196322858333588, + "learning_rate": 4.99472799961295e-05, + "loss": 0.0005, + "step": 405 + }, + { + "epoch": 0.2847624057513589, + "grad_norm": 0.6420115828514099, + "learning_rate": 4.994633470314387e-05, + "loss": 0.0026, + "step": 406 + }, + { + "epoch": 0.2854637909871997, + "grad_norm": 213.8779296875, + "learning_rate": 4.994538101963958e-05, + "loss": 2.1989, + "step": 407 + }, + { + "epoch": 0.2861651762230405, + "grad_norm": 108.08120727539062, + "learning_rate": 4.994441894593741e-05, + "loss": 3.9965, + "step": 408 + }, + { + "epoch": 0.2868665614588813, + "grad_norm": 0.40015965700149536, + "learning_rate": 4.994344848236094e-05, + "loss": 0.0016, + "step": 409 + }, + { + "epoch": 0.2875679466947221, + "grad_norm": 125.80431365966797, + "learning_rate": 4.9942469629236585e-05, + "loss": 2.3139, + "step": 410 + }, + { + "epoch": 0.28826933193056287, + "grad_norm": 101.27831268310547, + "learning_rate": 4.994148238689357e-05, + "loss": 2.1065, + "step": 411 + }, + { + "epoch": 0.28897071716640366, + "grad_norm": 3.526338577270508, + "learning_rate": 4.9940486755663946e-05, + "loss": 0.015, + "step": 412 + }, + { + "epoch": 0.28967210240224445, + "grad_norm": 2.501554489135742, + "learning_rate": 4.993948273588259e-05, + "loss": 0.0123, + "step": 413 + }, + { + "epoch": 0.29037348763808524, + "grad_norm": 75.39434814453125, + "learning_rate": 4.993847032788719e-05, + "loss": 0.7107, + "step": 414 + }, + { + "epoch": 0.291074872873926, + "grad_norm": 3.751978635787964, + "learning_rate": 4.993744953201827e-05, + "loss": 0.0183, + "step": 415 + }, + { + "epoch": 0.29177625810976676, + "grad_norm": 5.31666374206543, + "learning_rate": 4.9936420348619164e-05, + "loss": 0.0385, + "step": 416 + }, + { + "epoch": 0.29247764334560755, + "grad_norm": 104.77239227294922, + "learning_rate": 4.9935382778036036e-05, + "loss": 1.7353, + "step": 417 + }, + { + "epoch": 0.29317902858144834, + "grad_norm": 92.44829559326172, + "learning_rate": 4.993433682061786e-05, + "loss": 1.1895, + "step": 418 + }, + { + "epoch": 0.29388041381728913, + "grad_norm": 89.00714874267578, + "learning_rate": 4.993328247671643e-05, + "loss": 1.2323, + "step": 419 + }, + { + "epoch": 0.2945817990531299, + "grad_norm": 172.867431640625, + "learning_rate": 4.993221974668637e-05, + "loss": 1.7303, + "step": 420 + }, + { + "epoch": 0.2952831842889707, + "grad_norm": 19.69775390625, + "learning_rate": 4.9931148630885127e-05, + "loss": 0.1678, + "step": 421 + }, + { + "epoch": 0.2959845695248115, + "grad_norm": 22.828763961791992, + "learning_rate": 4.9930069129672965e-05, + "loss": 0.3184, + "step": 422 + }, + { + "epoch": 0.2966859547606523, + "grad_norm": 19.460153579711914, + "learning_rate": 4.992898124341296e-05, + "loss": 0.2748, + "step": 423 + }, + { + "epoch": 0.2973873399964931, + "grad_norm": 2.9695911407470703, + "learning_rate": 4.9927884972471015e-05, + "loss": 0.1925, + "step": 424 + }, + { + "epoch": 0.2980887252323339, + "grad_norm": 12.33647346496582, + "learning_rate": 4.9926780317215854e-05, + "loss": 0.1908, + "step": 425 + }, + { + "epoch": 0.29879011046817466, + "grad_norm": 2.8421826362609863, + "learning_rate": 4.992566727801902e-05, + "loss": 0.101, + "step": 426 + }, + { + "epoch": 0.29949149570401545, + "grad_norm": 171.8658447265625, + "learning_rate": 4.9924545855254875e-05, + "loss": 3.9311, + "step": 427 + }, + { + "epoch": 0.30019288093985624, + "grad_norm": 83.66468048095703, + "learning_rate": 4.992341604930061e-05, + "loss": 1.6806, + "step": 428 + }, + { + "epoch": 0.300894266175697, + "grad_norm": 0.7609736323356628, + "learning_rate": 4.9922277860536216e-05, + "loss": 0.0041, + "step": 429 + }, + { + "epoch": 0.30159565141153777, + "grad_norm": 4.077641487121582, + "learning_rate": 4.9921131289344514e-05, + "loss": 0.0418, + "step": 430 + }, + { + "epoch": 0.30229703664737856, + "grad_norm": 3.2341866493225098, + "learning_rate": 4.9919976336111155e-05, + "loss": 0.0293, + "step": 431 + }, + { + "epoch": 0.30299842188321935, + "grad_norm": 148.74095153808594, + "learning_rate": 4.991881300122458e-05, + "loss": 3.3951, + "step": 432 + }, + { + "epoch": 0.30369980711906014, + "grad_norm": 77.130126953125, + "learning_rate": 4.99176412850761e-05, + "loss": 2.9407, + "step": 433 + }, + { + "epoch": 0.30440119235490093, + "grad_norm": 162.98777770996094, + "learning_rate": 4.991646118805978e-05, + "loss": 4.3504, + "step": 434 + }, + { + "epoch": 0.3051025775907417, + "grad_norm": 84.4205551147461, + "learning_rate": 4.991527271057256e-05, + "loss": 1.7142, + "step": 435 + }, + { + "epoch": 0.3058039628265825, + "grad_norm": 0.8636782765388489, + "learning_rate": 4.991407585301416e-05, + "loss": 0.0053, + "step": 436 + }, + { + "epoch": 0.3065053480624233, + "grad_norm": 1.0418014526367188, + "learning_rate": 4.9912870615787144e-05, + "loss": 0.0059, + "step": 437 + }, + { + "epoch": 0.3072067332982641, + "grad_norm": 0.7050487399101257, + "learning_rate": 4.991165699929688e-05, + "loss": 0.0041, + "step": 438 + }, + { + "epoch": 0.3079081185341049, + "grad_norm": 0.3087022304534912, + "learning_rate": 4.991043500395157e-05, + "loss": 0.0017, + "step": 439 + }, + { + "epoch": 0.30860950376994567, + "grad_norm": 0.09886661171913147, + "learning_rate": 4.990920463016221e-05, + "loss": 0.0005, + "step": 440 + }, + { + "epoch": 0.3093108890057864, + "grad_norm": 0.03563763573765755, + "learning_rate": 4.9907965878342634e-05, + "loss": 0.0002, + "step": 441 + }, + { + "epoch": 0.3100122742416272, + "grad_norm": 79.55757141113281, + "learning_rate": 4.990671874890948e-05, + "loss": 2.966, + "step": 442 + }, + { + "epoch": 0.310713659477468, + "grad_norm": 0.6562897562980652, + "learning_rate": 4.990546324228222e-05, + "loss": 0.0022, + "step": 443 + }, + { + "epoch": 0.3114150447133088, + "grad_norm": 74.84564208984375, + "learning_rate": 4.990419935888313e-05, + "loss": 3.0154, + "step": 444 + }, + { + "epoch": 0.31211642994914957, + "grad_norm": 0.03038967400789261, + "learning_rate": 4.9902927099137305e-05, + "loss": 0.0002, + "step": 445 + }, + { + "epoch": 0.31281781518499036, + "grad_norm": 0.023526914417743683, + "learning_rate": 4.9901646463472674e-05, + "loss": 0.0001, + "step": 446 + }, + { + "epoch": 0.31351920042083115, + "grad_norm": 0.08529358357191086, + "learning_rate": 4.990035745231995e-05, + "loss": 0.0005, + "step": 447 + }, + { + "epoch": 0.31422058565667194, + "grad_norm": 0.11940905451774597, + "learning_rate": 4.989906006611269e-05, + "loss": 0.0007, + "step": 448 + }, + { + "epoch": 0.3149219708925127, + "grad_norm": 99.68978118896484, + "learning_rate": 4.989775430528727e-05, + "loss": 1.6958, + "step": 449 + }, + { + "epoch": 0.3156233561283535, + "grad_norm": 0.3344658315181732, + "learning_rate": 4.9896440170282864e-05, + "loss": 0.002, + "step": 450 + }, + { + "epoch": 0.3163247413641943, + "grad_norm": 0.7184603810310364, + "learning_rate": 4.9895117661541476e-05, + "loss": 0.0043, + "step": 451 + }, + { + "epoch": 0.3170261266000351, + "grad_norm": 0.5113491415977478, + "learning_rate": 4.989378677950791e-05, + "loss": 0.0031, + "step": 452 + }, + { + "epoch": 0.31772751183587583, + "grad_norm": 0.49223193526268005, + "learning_rate": 4.989244752462982e-05, + "loss": 0.0029, + "step": 453 + }, + { + "epoch": 0.3184288970717166, + "grad_norm": 64.00535583496094, + "learning_rate": 4.989109989735763e-05, + "loss": 1.2629, + "step": 454 + }, + { + "epoch": 0.3191302823075574, + "grad_norm": 0.18930155038833618, + "learning_rate": 4.988974389814463e-05, + "loss": 0.0011, + "step": 455 + }, + { + "epoch": 0.3198316675433982, + "grad_norm": 0.3088482916355133, + "learning_rate": 4.9888379527446885e-05, + "loss": 0.0018, + "step": 456 + }, + { + "epoch": 0.320533052779239, + "grad_norm": 71.42636108398438, + "learning_rate": 4.988700678572329e-05, + "loss": 1.9691, + "step": 457 + }, + { + "epoch": 0.3212344380150798, + "grad_norm": 0.20487411320209503, + "learning_rate": 4.988562567343557e-05, + "loss": 0.0012, + "step": 458 + }, + { + "epoch": 0.32193582325092057, + "grad_norm": 73.37615203857422, + "learning_rate": 4.988423619104823e-05, + "loss": 1.8636, + "step": 459 + }, + { + "epoch": 0.32263720848676136, + "grad_norm": 66.4114761352539, + "learning_rate": 4.988283833902863e-05, + "loss": 1.3073, + "step": 460 + }, + { + "epoch": 0.32333859372260215, + "grad_norm": 4.2518744468688965, + "learning_rate": 4.988143211784693e-05, + "loss": 0.0318, + "step": 461 + }, + { + "epoch": 0.32403997895844294, + "grad_norm": 4.782449245452881, + "learning_rate": 4.9880017527976083e-05, + "loss": 0.0384, + "step": 462 + }, + { + "epoch": 0.32474136419428373, + "grad_norm": 22.517131805419922, + "learning_rate": 4.9878594569891896e-05, + "loss": 0.2973, + "step": 463 + }, + { + "epoch": 0.3254427494301245, + "grad_norm": 5.63627290725708, + "learning_rate": 4.987716324407296e-05, + "loss": 0.0496, + "step": 464 + }, + { + "epoch": 0.32614413466596526, + "grad_norm": 6.364257335662842, + "learning_rate": 4.9875723551000696e-05, + "loss": 0.0578, + "step": 465 + }, + { + "epoch": 0.32684551990180605, + "grad_norm": 147.82330322265625, + "learning_rate": 4.987427549115933e-05, + "loss": 1.8188, + "step": 466 + }, + { + "epoch": 0.32754690513764684, + "grad_norm": 3.7355575561523438, + "learning_rate": 4.987281906503591e-05, + "loss": 0.0274, + "step": 467 + }, + { + "epoch": 0.32824829037348763, + "grad_norm": 0.6107962131500244, + "learning_rate": 4.987135427312029e-05, + "loss": 0.0038, + "step": 468 + }, + { + "epoch": 0.3289496756093284, + "grad_norm": 61.26199722290039, + "learning_rate": 4.9869881115905146e-05, + "loss": 0.8915, + "step": 469 + }, + { + "epoch": 0.3296510608451692, + "grad_norm": 0.3030995726585388, + "learning_rate": 4.986839959388596e-05, + "loss": 0.0018, + "step": 470 + }, + { + "epoch": 0.33035244608101, + "grad_norm": 0.12271957844495773, + "learning_rate": 4.986690970756104e-05, + "loss": 0.0007, + "step": 471 + }, + { + "epoch": 0.3310538313168508, + "grad_norm": 71.7486801147461, + "learning_rate": 4.9865411457431485e-05, + "loss": 1.8673, + "step": 472 + }, + { + "epoch": 0.3317552165526916, + "grad_norm": 69.17851257324219, + "learning_rate": 4.986390484400124e-05, + "loss": 1.6587, + "step": 473 + }, + { + "epoch": 0.33245660178853237, + "grad_norm": 0.18208004534244537, + "learning_rate": 4.9862389867777023e-05, + "loss": 0.001, + "step": 474 + }, + { + "epoch": 0.33315798702437316, + "grad_norm": 122.65370178222656, + "learning_rate": 4.986086652926839e-05, + "loss": 0.9068, + "step": 475 + }, + { + "epoch": 0.33385937226021395, + "grad_norm": 0.6119470596313477, + "learning_rate": 4.9859334828987715e-05, + "loss": 0.0039, + "step": 476 + }, + { + "epoch": 0.3345607574960547, + "grad_norm": 4.483019828796387, + "learning_rate": 4.985779476745017e-05, + "loss": 0.0343, + "step": 477 + }, + { + "epoch": 0.3352621427318955, + "grad_norm": 0.7592105269432068, + "learning_rate": 4.985624634517374e-05, + "loss": 0.005, + "step": 478 + }, + { + "epoch": 0.33596352796773626, + "grad_norm": 0.6566614508628845, + "learning_rate": 4.985468956267922e-05, + "loss": 0.0037, + "step": 479 + }, + { + "epoch": 0.33666491320357705, + "grad_norm": 67.39823150634766, + "learning_rate": 4.985312442049023e-05, + "loss": 1.7049, + "step": 480 + }, + { + "epoch": 0.33736629843941784, + "grad_norm": 0.13803933560848236, + "learning_rate": 4.98515509191332e-05, + "loss": 0.0008, + "step": 481 + }, + { + "epoch": 0.33806768367525863, + "grad_norm": 72.15523529052734, + "learning_rate": 4.984996905913736e-05, + "loss": 1.6634, + "step": 482 + }, + { + "epoch": 0.3387690689110994, + "grad_norm": 0.10033612698316574, + "learning_rate": 4.984837884103475e-05, + "loss": 0.0005, + "step": 483 + }, + { + "epoch": 0.3394704541469402, + "grad_norm": 62.519317626953125, + "learning_rate": 4.984678026536024e-05, + "loss": 1.2577, + "step": 484 + }, + { + "epoch": 0.340171839382781, + "grad_norm": 0.36229056119918823, + "learning_rate": 4.984517333265148e-05, + "loss": 0.0023, + "step": 485 + }, + { + "epoch": 0.3408732246186218, + "grad_norm": 0.8776547312736511, + "learning_rate": 4.9843558043448985e-05, + "loss": 0.0056, + "step": 486 + }, + { + "epoch": 0.3415746098544626, + "grad_norm": 108.73171997070312, + "learning_rate": 4.984193439829602e-05, + "loss": 2.7267, + "step": 487 + }, + { + "epoch": 0.3422759950903034, + "grad_norm": 0.6344736814498901, + "learning_rate": 4.984030239773868e-05, + "loss": 0.0041, + "step": 488 + }, + { + "epoch": 0.3429773803261441, + "grad_norm": 113.62140655517578, + "learning_rate": 4.98386620423259e-05, + "loss": 1.7271, + "step": 489 + }, + { + "epoch": 0.3436787655619849, + "grad_norm": 1.618384599685669, + "learning_rate": 4.983701333260938e-05, + "loss": 0.011, + "step": 490 + }, + { + "epoch": 0.3443801507978257, + "grad_norm": 138.78854370117188, + "learning_rate": 4.9835356269143676e-05, + "loss": 1.9143, + "step": 491 + }, + { + "epoch": 0.3450815360336665, + "grad_norm": 49.62142562866211, + "learning_rate": 4.98336908524861e-05, + "loss": 0.753, + "step": 492 + }, + { + "epoch": 0.34578292126950727, + "grad_norm": 3.6228299140930176, + "learning_rate": 4.9832017083196816e-05, + "loss": 0.0277, + "step": 493 + }, + { + "epoch": 0.34648430650534806, + "grad_norm": 24.915790557861328, + "learning_rate": 4.98303349618388e-05, + "loss": 0.356, + "step": 494 + }, + { + "epoch": 0.34718569174118885, + "grad_norm": 7.880053997039795, + "learning_rate": 4.9828644488977795e-05, + "loss": 0.0727, + "step": 495 + }, + { + "epoch": 0.34788707697702964, + "grad_norm": 12.95199203491211, + "learning_rate": 4.982694566518239e-05, + "loss": 0.25, + "step": 496 + }, + { + "epoch": 0.34858846221287043, + "grad_norm": 2.059384822845459, + "learning_rate": 4.982523849102397e-05, + "loss": 0.2069, + "step": 497 + }, + { + "epoch": 0.3492898474487112, + "grad_norm": 17.132720947265625, + "learning_rate": 4.982352296707675e-05, + "loss": 0.2093, + "step": 498 + }, + { + "epoch": 0.349991232684552, + "grad_norm": 12.591998100280762, + "learning_rate": 4.98217990939177e-05, + "loss": 0.1889, + "step": 499 + }, + { + "epoch": 0.3506926179203928, + "grad_norm": 12.747396469116211, + "learning_rate": 4.982006687212666e-05, + "loss": 0.1755, + "step": 500 + }, + { + "epoch": 0.35139400315623354, + "grad_norm": 77.32592010498047, + "learning_rate": 4.981832630228624e-05, + "loss": 0.9517, + "step": 501 + }, + { + "epoch": 0.3520953883920743, + "grad_norm": 8.452153205871582, + "learning_rate": 4.9816577384981866e-05, + "loss": 0.0913, + "step": 502 + }, + { + "epoch": 0.3527967736279151, + "grad_norm": 2.1267127990722656, + "learning_rate": 4.9814820120801776e-05, + "loss": 0.0178, + "step": 503 + }, + { + "epoch": 0.3534981588637559, + "grad_norm": 0.8161488771438599, + "learning_rate": 4.981305451033701e-05, + "loss": 0.0058, + "step": 504 + }, + { + "epoch": 0.3541995440995967, + "grad_norm": 0.10489612072706223, + "learning_rate": 4.981128055418143e-05, + "loss": 0.0006, + "step": 505 + }, + { + "epoch": 0.3549009293354375, + "grad_norm": 69.51275634765625, + "learning_rate": 4.9809498252931674e-05, + "loss": 2.62, + "step": 506 + }, + { + "epoch": 0.3556023145712783, + "grad_norm": 0.004451240878552198, + "learning_rate": 4.980770760718721e-05, + "loss": 0.0, + "step": 507 + }, + { + "epoch": 0.35630369980711907, + "grad_norm": 67.6517105102539, + "learning_rate": 4.9805908617550335e-05, + "loss": 3.25, + "step": 508 + }, + { + "epoch": 0.35700508504295986, + "grad_norm": 0.016789492219686508, + "learning_rate": 4.980410128462609e-05, + "loss": 0.0001, + "step": 509 + }, + { + "epoch": 0.35770647027880065, + "grad_norm": 0.025487884879112244, + "learning_rate": 4.980228560902239e-05, + "loss": 0.0001, + "step": 510 + }, + { + "epoch": 0.35840785551464144, + "grad_norm": 0.042408283799886703, + "learning_rate": 4.98004615913499e-05, + "loss": 0.0002, + "step": 511 + }, + { + "epoch": 0.35910924075048223, + "grad_norm": 0.038653161376714706, + "learning_rate": 4.979862923222212e-05, + "loss": 0.0002, + "step": 512 + }, + { + "epoch": 0.35981062598632296, + "grad_norm": 0.03510050103068352, + "learning_rate": 4.9796788532255356e-05, + "loss": 0.0002, + "step": 513 + }, + { + "epoch": 0.36051201122216375, + "grad_norm": 0.008441315963864326, + "learning_rate": 4.979493949206872e-05, + "loss": 0.0, + "step": 514 + }, + { + "epoch": 0.36121339645800454, + "grad_norm": 0.006373463664203882, + "learning_rate": 4.979308211228411e-05, + "loss": 0.0, + "step": 515 + }, + { + "epoch": 0.36191478169384533, + "grad_norm": 79.59523010253906, + "learning_rate": 4.979121639352625e-05, + "loss": 2.9487, + "step": 516 + }, + { + "epoch": 0.3626161669296861, + "grad_norm": 153.0602264404297, + "learning_rate": 4.9789342336422665e-05, + "loss": 3.6533, + "step": 517 + }, + { + "epoch": 0.3633175521655269, + "grad_norm": 0.7085576057434082, + "learning_rate": 4.9787459941603666e-05, + "loss": 0.0042, + "step": 518 + }, + { + "epoch": 0.3640189374013677, + "grad_norm": 155.89154052734375, + "learning_rate": 4.9785569209702395e-05, + "loss": 7.9018, + "step": 519 + }, + { + "epoch": 0.3647203226372085, + "grad_norm": 5.253132343292236, + "learning_rate": 4.9783670141354776e-05, + "loss": 0.042, + "step": 520 + }, + { + "epoch": 0.3654217078730493, + "grad_norm": 26.310007095336914, + "learning_rate": 4.978176273719957e-05, + "loss": 0.2762, + "step": 521 + }, + { + "epoch": 0.3661230931088901, + "grad_norm": 9.570679664611816, + "learning_rate": 4.977984699787829e-05, + "loss": 0.0898, + "step": 522 + }, + { + "epoch": 0.36682447834473086, + "grad_norm": 214.85060119628906, + "learning_rate": 4.977792292403529e-05, + "loss": 6.7239, + "step": 523 + }, + { + "epoch": 0.36752586358057165, + "grad_norm": 2.7513771057128906, + "learning_rate": 4.977599051631773e-05, + "loss": 0.0228, + "step": 524 + }, + { + "epoch": 0.3682272488164124, + "grad_norm": 33.629356384277344, + "learning_rate": 4.977404977537554e-05, + "loss": 0.4374, + "step": 525 + }, + { + "epoch": 0.3689286340522532, + "grad_norm": 3.8864808082580566, + "learning_rate": 4.97721007018615e-05, + "loss": 0.0308, + "step": 526 + }, + { + "epoch": 0.36963001928809397, + "grad_norm": 1.0654228925704956, + "learning_rate": 4.9770143296431144e-05, + "loss": 0.0077, + "step": 527 + }, + { + "epoch": 0.37033140452393476, + "grad_norm": 131.74990844726562, + "learning_rate": 4.9768177559742844e-05, + "loss": 4.0998, + "step": 528 + }, + { + "epoch": 0.37103278975977555, + "grad_norm": 0.10420957952737808, + "learning_rate": 4.976620349245776e-05, + "loss": 0.0006, + "step": 529 + }, + { + "epoch": 0.37173417499561634, + "grad_norm": 72.78388977050781, + "learning_rate": 4.976422109523985e-05, + "loss": 3.8242, + "step": 530 + }, + { + "epoch": 0.37243556023145713, + "grad_norm": 69.0342025756836, + "learning_rate": 4.976223036875588e-05, + "loss": 1.9872, + "step": 531 + }, + { + "epoch": 0.3731369454672979, + "grad_norm": 0.43323075771331787, + "learning_rate": 4.9760231313675426e-05, + "loss": 0.0028, + "step": 532 + }, + { + "epoch": 0.3738383307031387, + "grad_norm": 1.0747731924057007, + "learning_rate": 4.975822393067084e-05, + "loss": 0.0061, + "step": 533 + }, + { + "epoch": 0.3745397159389795, + "grad_norm": 69.71436309814453, + "learning_rate": 4.9756208220417304e-05, + "loss": 1.8796, + "step": 534 + }, + { + "epoch": 0.3752411011748203, + "grad_norm": 127.05543518066406, + "learning_rate": 4.975418418359278e-05, + "loss": 2.8441, + "step": 535 + }, + { + "epoch": 0.3759424864106611, + "grad_norm": 0.9959923624992371, + "learning_rate": 4.975215182087804e-05, + "loss": 0.0065, + "step": 536 + }, + { + "epoch": 0.3766438716465018, + "grad_norm": 2.1746952533721924, + "learning_rate": 4.9750111132956654e-05, + "loss": 0.0156, + "step": 537 + }, + { + "epoch": 0.3773452568823426, + "grad_norm": 92.57381439208984, + "learning_rate": 4.9748062120514996e-05, + "loss": 1.4085, + "step": 538 + }, + { + "epoch": 0.3780466421181834, + "grad_norm": 75.45511627197266, + "learning_rate": 4.974600478424223e-05, + "loss": 0.9909, + "step": 539 + }, + { + "epoch": 0.3787480273540242, + "grad_norm": 25.877460479736328, + "learning_rate": 4.974393912483034e-05, + "loss": 0.3578, + "step": 540 + }, + { + "epoch": 0.379449412589865, + "grad_norm": 12.262901306152344, + "learning_rate": 4.9741865142974076e-05, + "loss": 0.1338, + "step": 541 + }, + { + "epoch": 0.38015079782570577, + "grad_norm": 10.54548168182373, + "learning_rate": 4.973978283937102e-05, + "loss": 0.2111, + "step": 542 + }, + { + "epoch": 0.38085218306154656, + "grad_norm": 35.909976959228516, + "learning_rate": 4.973769221472153e-05, + "loss": 0.5549, + "step": 543 + }, + { + "epoch": 0.38155356829738735, + "grad_norm": 13.012696266174316, + "learning_rate": 4.973559326972878e-05, + "loss": 0.1784, + "step": 544 + }, + { + "epoch": 0.38225495353322814, + "grad_norm": 16.022974014282227, + "learning_rate": 4.973348600509874e-05, + "loss": 0.2235, + "step": 545 + }, + { + "epoch": 0.3829563387690689, + "grad_norm": 22.86570167541504, + "learning_rate": 4.973137042154016e-05, + "loss": 0.3561, + "step": 546 + }, + { + "epoch": 0.3836577240049097, + "grad_norm": 13.543822288513184, + "learning_rate": 4.972924651976461e-05, + "loss": 0.3077, + "step": 547 + }, + { + "epoch": 0.3843591092407505, + "grad_norm": 17.42881965637207, + "learning_rate": 4.972711430048644e-05, + "loss": 0.2779, + "step": 548 + }, + { + "epoch": 0.38506049447659124, + "grad_norm": 19.7313175201416, + "learning_rate": 4.972497376442282e-05, + "loss": 0.2582, + "step": 549 + }, + { + "epoch": 0.38576187971243203, + "grad_norm": 12.354894638061523, + "learning_rate": 4.9722824912293695e-05, + "loss": 0.2374, + "step": 550 + }, + { + "epoch": 0.3864632649482728, + "grad_norm": 17.16146469116211, + "learning_rate": 4.972066774482182e-05, + "loss": 0.2085, + "step": 551 + }, + { + "epoch": 0.3871646501841136, + "grad_norm": 49.288734436035156, + "learning_rate": 4.971850226273274e-05, + "loss": 0.6211, + "step": 552 + }, + { + "epoch": 0.3878660354199544, + "grad_norm": 4.870689392089844, + "learning_rate": 4.97163284667548e-05, + "loss": 0.0413, + "step": 553 + }, + { + "epoch": 0.3885674206557952, + "grad_norm": 4.268405437469482, + "learning_rate": 4.971414635761914e-05, + "loss": 0.0354, + "step": 554 + }, + { + "epoch": 0.389268805891636, + "grad_norm": 2.800025701522827, + "learning_rate": 4.971195593605969e-05, + "loss": 0.0215, + "step": 555 + }, + { + "epoch": 0.3899701911274768, + "grad_norm": 136.0956573486328, + "learning_rate": 4.9709757202813195e-05, + "loss": 3.3931, + "step": 556 + }, + { + "epoch": 0.39067157636331756, + "grad_norm": 68.56623840332031, + "learning_rate": 4.970755015861918e-05, + "loss": 1.8678, + "step": 557 + }, + { + "epoch": 0.39137296159915835, + "grad_norm": 68.25215911865234, + "learning_rate": 4.970533480421996e-05, + "loss": 1.9984, + "step": 558 + }, + { + "epoch": 0.39207434683499914, + "grad_norm": 69.02233123779297, + "learning_rate": 4.9703111140360656e-05, + "loss": 2.0662, + "step": 559 + }, + { + "epoch": 0.39277573207083993, + "grad_norm": 61.60824203491211, + "learning_rate": 4.970087916778918e-05, + "loss": 1.5281, + "step": 560 + }, + { + "epoch": 0.39347711730668067, + "grad_norm": 48.50178146362305, + "learning_rate": 4.9698638887256254e-05, + "loss": 0.8115, + "step": 561 + }, + { + "epoch": 0.39417850254252146, + "grad_norm": 41.32649612426758, + "learning_rate": 4.969639029951537e-05, + "loss": 0.6374, + "step": 562 + }, + { + "epoch": 0.39487988777836225, + "grad_norm": 6.946910858154297, + "learning_rate": 4.9694133405322815e-05, + "loss": 0.2402, + "step": 563 + }, + { + "epoch": 0.39558127301420304, + "grad_norm": 22.8358097076416, + "learning_rate": 4.9691868205437694e-05, + "loss": 0.3883, + "step": 564 + }, + { + "epoch": 0.39628265825004383, + "grad_norm": 15.402759552001953, + "learning_rate": 4.968959470062189e-05, + "loss": 0.7708, + "step": 565 + }, + { + "epoch": 0.3969840434858846, + "grad_norm": 28.08781623840332, + "learning_rate": 4.968731289164006e-05, + "loss": 0.7703, + "step": 566 + }, + { + "epoch": 0.3976854287217254, + "grad_norm": 26.910886764526367, + "learning_rate": 4.9685022779259696e-05, + "loss": 0.6299, + "step": 567 + }, + { + "epoch": 0.3983868139575662, + "grad_norm": 24.341957092285156, + "learning_rate": 4.968272436425105e-05, + "loss": 0.5751, + "step": 568 + }, + { + "epoch": 0.399088199193407, + "grad_norm": 23.469263076782227, + "learning_rate": 4.9680417647387186e-05, + "loss": 0.4664, + "step": 569 + }, + { + "epoch": 0.3997895844292478, + "grad_norm": 15.347972869873047, + "learning_rate": 4.967810262944394e-05, + "loss": 0.2252, + "step": 570 + }, + { + "epoch": 0.40049096966508857, + "grad_norm": 5.6460280418396, + "learning_rate": 4.967577931119996e-05, + "loss": 0.2232, + "step": 571 + }, + { + "epoch": 0.40119235490092936, + "grad_norm": 4.2383856773376465, + "learning_rate": 4.967344769343667e-05, + "loss": 0.0396, + "step": 572 + }, + { + "epoch": 0.4018937401367701, + "grad_norm": 1.6932095289230347, + "learning_rate": 4.96711077769383e-05, + "loss": 0.0137, + "step": 573 + }, + { + "epoch": 0.4025951253726109, + "grad_norm": 0.41152113676071167, + "learning_rate": 4.966875956249186e-05, + "loss": 0.0029, + "step": 574 + }, + { + "epoch": 0.4032965106084517, + "grad_norm": 0.06895897537469864, + "learning_rate": 4.9666403050887155e-05, + "loss": 0.0004, + "step": 575 + }, + { + "epoch": 0.40399789584429247, + "grad_norm": 0.004564318805932999, + "learning_rate": 4.9664038242916786e-05, + "loss": 0.0, + "step": 576 + }, + { + "epoch": 0.40469928108013326, + "grad_norm": 0.0019707169849425554, + "learning_rate": 4.966166513937613e-05, + "loss": 0.0, + "step": 577 + }, + { + "epoch": 0.40540066631597405, + "grad_norm": 0.00037522497586905956, + "learning_rate": 4.965928374106337e-05, + "loss": 0.0, + "step": 578 + }, + { + "epoch": 0.40610205155181484, + "grad_norm": 5.6252967624459416e-05, + "learning_rate": 4.965689404877946e-05, + "loss": 0.0, + "step": 579 + }, + { + "epoch": 0.4068034367876556, + "grad_norm": 1.1128309779451229e-05, + "learning_rate": 4.9654496063328174e-05, + "loss": 0.0, + "step": 580 + }, + { + "epoch": 0.4075048220234964, + "grad_norm": 119.89021301269531, + "learning_rate": 4.965208978551605e-05, + "loss": 10.8344, + "step": 581 + }, + { + "epoch": 0.4082062072593372, + "grad_norm": 1.3851165931555443e-05, + "learning_rate": 4.964967521615241e-05, + "loss": 0.0, + "step": 582 + }, + { + "epoch": 0.408907592495178, + "grad_norm": 3.7249556044116616e-06, + "learning_rate": 4.96472523560494e-05, + "loss": 0.0, + "step": 583 + }, + { + "epoch": 0.4096089777310188, + "grad_norm": 2.917176971095614e-06, + "learning_rate": 4.9644821206021906e-05, + "loss": 0.0, + "step": 584 + }, + { + "epoch": 0.4103103629668595, + "grad_norm": 3.3871406230900902e-06, + "learning_rate": 4.964238176688765e-05, + "loss": 0.0, + "step": 585 + }, + { + "epoch": 0.4110117482027003, + "grad_norm": 2.260056135128252e-06, + "learning_rate": 4.9639934039467106e-05, + "loss": 0.0, + "step": 586 + }, + { + "epoch": 0.4117131334385411, + "grad_norm": 1.5038928040667088e-06, + "learning_rate": 4.963747802458355e-05, + "loss": 0.0, + "step": 587 + }, + { + "epoch": 0.4124145186743819, + "grad_norm": 59.293521881103516, + "learning_rate": 4.963501372306306e-05, + "loss": 5.3994, + "step": 588 + }, + { + "epoch": 0.4131159039102227, + "grad_norm": 59.763816833496094, + "learning_rate": 4.963254113573446e-05, + "loss": 5.6505, + "step": 589 + }, + { + "epoch": 0.41381728914606347, + "grad_norm": 132.7192840576172, + "learning_rate": 4.963006026342941e-05, + "loss": 9.1499, + "step": 590 + }, + { + "epoch": 0.41451867438190426, + "grad_norm": 0.00024285895051434636, + "learning_rate": 4.962757110698232e-05, + "loss": 0.0, + "step": 591 + }, + { + "epoch": 0.41522005961774505, + "grad_norm": 0.0009618580807000399, + "learning_rate": 4.9625073667230404e-05, + "loss": 0.0, + "step": 592 + }, + { + "epoch": 0.41592144485358584, + "grad_norm": 0.004245178773999214, + "learning_rate": 4.962256794501365e-05, + "loss": 0.0, + "step": 593 + }, + { + "epoch": 0.41662283008942663, + "grad_norm": 0.01505737192928791, + "learning_rate": 4.962005394117486e-05, + "loss": 0.0001, + "step": 594 + }, + { + "epoch": 0.4173242153252674, + "grad_norm": 0.01870640181005001, + "learning_rate": 4.961753165655958e-05, + "loss": 0.0001, + "step": 595 + }, + { + "epoch": 0.4180256005611082, + "grad_norm": 0.047609515488147736, + "learning_rate": 4.961500109201618e-05, + "loss": 0.0003, + "step": 596 + }, + { + "epoch": 0.41872698579694895, + "grad_norm": 0.15332278609275818, + "learning_rate": 4.9612462248395775e-05, + "loss": 0.001, + "step": 597 + }, + { + "epoch": 0.41942837103278974, + "grad_norm": 113.53760528564453, + "learning_rate": 4.96099151265523e-05, + "loss": 2.3303, + "step": 598 + }, + { + "epoch": 0.4201297562686305, + "grad_norm": 0.6282661557197571, + "learning_rate": 4.9607359727342465e-05, + "loss": 0.0044, + "step": 599 + }, + { + "epoch": 0.4208311415044713, + "grad_norm": 43.630615234375, + "learning_rate": 4.9604796051625756e-05, + "loss": 0.6733, + "step": 600 + }, + { + "epoch": 0.4215325267403121, + "grad_norm": 72.0724868774414, + "learning_rate": 4.960222410026444e-05, + "loss": 1.4145, + "step": 601 + }, + { + "epoch": 0.4222339119761529, + "grad_norm": 6.166507244110107, + "learning_rate": 4.959964387412358e-05, + "loss": 0.0608, + "step": 602 + }, + { + "epoch": 0.4229352972119937, + "grad_norm": 80.03913879394531, + "learning_rate": 4.9597055374071024e-05, + "loss": 1.2555, + "step": 603 + }, + { + "epoch": 0.4236366824478345, + "grad_norm": 17.456140518188477, + "learning_rate": 4.959445860097738e-05, + "loss": 0.238, + "step": 604 + }, + { + "epoch": 0.42433806768367527, + "grad_norm": 36.63374328613281, + "learning_rate": 4.9591853555716064e-05, + "loss": 0.8324, + "step": 605 + }, + { + "epoch": 0.42503945291951606, + "grad_norm": 16.27304458618164, + "learning_rate": 4.9589240239163255e-05, + "loss": 0.2125, + "step": 606 + }, + { + "epoch": 0.42574083815535685, + "grad_norm": 19.427776336669922, + "learning_rate": 4.9586618652197934e-05, + "loss": 0.2701, + "step": 607 + }, + { + "epoch": 0.42644222339119764, + "grad_norm": 8.516806602478027, + "learning_rate": 4.958398879570185e-05, + "loss": 0.3408, + "step": 608 + }, + { + "epoch": 0.4271436086270384, + "grad_norm": 1.857074499130249, + "learning_rate": 4.9581350670559524e-05, + "loss": 0.2039, + "step": 609 + }, + { + "epoch": 0.42784499386287916, + "grad_norm": 8.925201416015625, + "learning_rate": 4.9578704277658286e-05, + "loss": 0.09, + "step": 610 + }, + { + "epoch": 0.42854637909871995, + "grad_norm": 4.490631580352783, + "learning_rate": 4.957604961788823e-05, + "loss": 0.0384, + "step": 611 + }, + { + "epoch": 0.42924776433456074, + "grad_norm": 0.866288959980011, + "learning_rate": 4.957338669214222e-05, + "loss": 0.0063, + "step": 612 + }, + { + "epoch": 0.42994914957040153, + "grad_norm": 0.6116040349006653, + "learning_rate": 4.9570715501315925e-05, + "loss": 0.0035, + "step": 613 + }, + { + "epoch": 0.4306505348062423, + "grad_norm": 0.014350208453834057, + "learning_rate": 4.956803604630777e-05, + "loss": 0.0001, + "step": 614 + }, + { + "epoch": 0.4313519200420831, + "grad_norm": 0.004978867247700691, + "learning_rate": 4.956534832801897e-05, + "loss": 0.0, + "step": 615 + }, + { + "epoch": 0.4320533052779239, + "grad_norm": 0.0008859955123625696, + "learning_rate": 4.9562652347353535e-05, + "loss": 0.0, + "step": 616 + }, + { + "epoch": 0.4327546905137647, + "grad_norm": 0.0007402606424875557, + "learning_rate": 4.955994810521822e-05, + "loss": 0.0, + "step": 617 + }, + { + "epoch": 0.4334560757496055, + "grad_norm": 60.98609161376953, + "learning_rate": 4.955723560252259e-05, + "loss": 4.6234, + "step": 618 + }, + { + "epoch": 0.4341574609854463, + "grad_norm": 60.63393783569336, + "learning_rate": 4.955451484017896e-05, + "loss": 4.669, + "step": 619 + }, + { + "epoch": 0.43485884622128707, + "grad_norm": 0.00016736971156205982, + "learning_rate": 4.955178581910246e-05, + "loss": 0.0, + "step": 620 + }, + { + "epoch": 0.4355602314571278, + "grad_norm": 0.00026081904070451856, + "learning_rate": 4.9549048540210963e-05, + "loss": 0.0, + "step": 621 + }, + { + "epoch": 0.4362616166929686, + "grad_norm": 0.0002771130239125341, + "learning_rate": 4.954630300442513e-05, + "loss": 0.0, + "step": 622 + }, + { + "epoch": 0.4369630019288094, + "grad_norm": 0.00024890206987038255, + "learning_rate": 4.954354921266841e-05, + "loss": 0.0, + "step": 623 + }, + { + "epoch": 0.43766438716465017, + "grad_norm": 0.0003540183824952692, + "learning_rate": 4.954078716586702e-05, + "loss": 0.0, + "step": 624 + }, + { + "epoch": 0.43836577240049096, + "grad_norm": 61.1258659362793, + "learning_rate": 4.9538016864949953e-05, + "loss": 4.1063, + "step": 625 + }, + { + "epoch": 0.43906715763633175, + "grad_norm": 61.82120895385742, + "learning_rate": 4.953523831084897e-05, + "loss": 3.2575, + "step": 626 + }, + { + "epoch": 0.43976854287217254, + "grad_norm": 0.009341489523649216, + "learning_rate": 4.9532451504498644e-05, + "loss": 0.0001, + "step": 627 + }, + { + "epoch": 0.44046992810801333, + "grad_norm": 126.25260925292969, + "learning_rate": 4.952965644683627e-05, + "loss": 4.8894, + "step": 628 + }, + { + "epoch": 0.4411713133438541, + "grad_norm": 64.0586166381836, + "learning_rate": 4.9526853138801954e-05, + "loss": 2.0939, + "step": 629 + }, + { + "epoch": 0.4418726985796949, + "grad_norm": 66.46090698242188, + "learning_rate": 4.952404158133857e-05, + "loss": 2.1672, + "step": 630 + }, + { + "epoch": 0.4425740838155357, + "grad_norm": 8.158260345458984, + "learning_rate": 4.952122177539177e-05, + "loss": 0.089, + "step": 631 + }, + { + "epoch": 0.4432754690513765, + "grad_norm": 8.28303337097168, + "learning_rate": 4.951839372190996e-05, + "loss": 0.1792, + "step": 632 + }, + { + "epoch": 0.4439768542872172, + "grad_norm": 6.670780658721924, + "learning_rate": 4.951555742184435e-05, + "loss": 0.3495, + "step": 633 + }, + { + "epoch": 0.444678239523058, + "grad_norm": 18.7882022857666, + "learning_rate": 4.95127128761489e-05, + "loss": 0.2761, + "step": 634 + }, + { + "epoch": 0.4453796247588988, + "grad_norm": 3.4713053703308105, + "learning_rate": 4.950986008578037e-05, + "loss": 0.2429, + "step": 635 + }, + { + "epoch": 0.4460810099947396, + "grad_norm": 8.415278434753418, + "learning_rate": 4.950699905169825e-05, + "loss": 0.0883, + "step": 636 + }, + { + "epoch": 0.4467823952305804, + "grad_norm": 5.512305736541748, + "learning_rate": 4.950412977486484e-05, + "loss": 0.0515, + "step": 637 + }, + { + "epoch": 0.4474837804664212, + "grad_norm": 1.983485460281372, + "learning_rate": 4.95012522562452e-05, + "loss": 0.0162, + "step": 638 + }, + { + "epoch": 0.44818516570226197, + "grad_norm": 0.5145743489265442, + "learning_rate": 4.949836649680717e-05, + "loss": 0.0038, + "step": 639 + }, + { + "epoch": 0.44888655093810276, + "grad_norm": 0.0464518703520298, + "learning_rate": 4.949547249752134e-05, + "loss": 0.0003, + "step": 640 + }, + { + "epoch": 0.44958793617394355, + "grad_norm": 63.47880935668945, + "learning_rate": 4.9492570259361094e-05, + "loss": 2.606, + "step": 641 + }, + { + "epoch": 0.45028932140978434, + "grad_norm": 63.044219970703125, + "learning_rate": 4.948965978330258e-05, + "loss": 3.3714, + "step": 642 + }, + { + "epoch": 0.45099070664562513, + "grad_norm": 61.190128326416016, + "learning_rate": 4.948674107032471e-05, + "loss": 3.3129, + "step": 643 + }, + { + "epoch": 0.4516920918814659, + "grad_norm": 59.03745651245117, + "learning_rate": 4.948381412140917e-05, + "loss": 2.3587, + "step": 644 + }, + { + "epoch": 0.4523934771173067, + "grad_norm": 61.125858306884766, + "learning_rate": 4.948087893754043e-05, + "loss": 2.7351, + "step": 645 + }, + { + "epoch": 0.45309486235314744, + "grad_norm": 60.69954299926758, + "learning_rate": 4.947793551970571e-05, + "loss": 2.2391, + "step": 646 + }, + { + "epoch": 0.45379624758898823, + "grad_norm": 56.91267395019531, + "learning_rate": 4.9474983868895e-05, + "loss": 1.5786, + "step": 647 + }, + { + "epoch": 0.454497632824829, + "grad_norm": 0.8762818574905396, + "learning_rate": 4.947202398610108e-05, + "loss": 0.0067, + "step": 648 + }, + { + "epoch": 0.4551990180606698, + "grad_norm": 2.4974522590637207, + "learning_rate": 4.9469055872319483e-05, + "loss": 0.0224, + "step": 649 + }, + { + "epoch": 0.4559004032965106, + "grad_norm": 22.98019790649414, + "learning_rate": 4.9466079528548505e-05, + "loss": 0.3665, + "step": 650 + }, + { + "epoch": 0.4566017885323514, + "grad_norm": 16.03438949584961, + "learning_rate": 4.9463094955789214e-05, + "loss": 0.315, + "step": 651 + }, + { + "epoch": 0.4573031737681922, + "grad_norm": 8.860062599182129, + "learning_rate": 4.946010215504546e-05, + "loss": 0.1067, + "step": 652 + }, + { + "epoch": 0.458004559004033, + "grad_norm": 5.521644115447998, + "learning_rate": 4.945710112732386e-05, + "loss": 0.2473, + "step": 653 + }, + { + "epoch": 0.45870594423987376, + "grad_norm": 15.16481876373291, + "learning_rate": 4.945409187363376e-05, + "loss": 0.2257, + "step": 654 + }, + { + "epoch": 0.45940732947571455, + "grad_norm": 15.714698791503906, + "learning_rate": 4.945107439498732e-05, + "loss": 0.2359, + "step": 655 + }, + { + "epoch": 0.46010871471155534, + "grad_norm": 6.4135589599609375, + "learning_rate": 4.944804869239944e-05, + "loss": 0.2336, + "step": 656 + }, + { + "epoch": 0.46081009994739613, + "grad_norm": 9.600846290588379, + "learning_rate": 4.94450147668878e-05, + "loss": 0.1174, + "step": 657 + }, + { + "epoch": 0.46151148518323687, + "grad_norm": 7.477170467376709, + "learning_rate": 4.944197261947283e-05, + "loss": 0.0857, + "step": 658 + }, + { + "epoch": 0.46221287041907766, + "grad_norm": 2.0777835845947266, + "learning_rate": 4.943892225117774e-05, + "loss": 0.0196, + "step": 659 + }, + { + "epoch": 0.46291425565491845, + "grad_norm": 41.072837829589844, + "learning_rate": 4.9435863663028495e-05, + "loss": 0.7781, + "step": 660 + }, + { + "epoch": 0.46361564089075924, + "grad_norm": 0.5752171277999878, + "learning_rate": 4.943279685605383e-05, + "loss": 0.0045, + "step": 661 + }, + { + "epoch": 0.46431702612660003, + "grad_norm": 51.6488151550293, + "learning_rate": 4.9429721831285246e-05, + "loss": 1.1945, + "step": 662 + }, + { + "epoch": 0.4650184113624408, + "grad_norm": 0.11998272687196732, + "learning_rate": 4.942663858975701e-05, + "loss": 0.0009, + "step": 663 + }, + { + "epoch": 0.4657197965982816, + "grad_norm": 57.40029525756836, + "learning_rate": 4.9423547132506135e-05, + "loss": 1.7823, + "step": 664 + }, + { + "epoch": 0.4664211818341224, + "grad_norm": 0.07954677194356918, + "learning_rate": 4.942044746057242e-05, + "loss": 0.0006, + "step": 665 + }, + { + "epoch": 0.4671225670699632, + "grad_norm": 77.24893951416016, + "learning_rate": 4.941733957499841e-05, + "loss": 3.3749, + "step": 666 + }, + { + "epoch": 0.467823952305804, + "grad_norm": 57.04423522949219, + "learning_rate": 4.9414223476829434e-05, + "loss": 1.6771, + "step": 667 + }, + { + "epoch": 0.46852533754164477, + "grad_norm": 0.14917835593223572, + "learning_rate": 4.941109916711355e-05, + "loss": 0.0011, + "step": 668 + }, + { + "epoch": 0.46922672277748556, + "grad_norm": 47.98805236816406, + "learning_rate": 4.940796664690161e-05, + "loss": 1.0673, + "step": 669 + }, + { + "epoch": 0.4699281080133263, + "grad_norm": 63.10405349731445, + "learning_rate": 4.940482591724721e-05, + "loss": 1.7249, + "step": 670 + }, + { + "epoch": 0.4706294932491671, + "grad_norm": 47.01327133178711, + "learning_rate": 4.940167697920671e-05, + "loss": 0.9642, + "step": 671 + }, + { + "epoch": 0.4713308784850079, + "grad_norm": 4.804792881011963, + "learning_rate": 4.9398519833839234e-05, + "loss": 0.0514, + "step": 672 + }, + { + "epoch": 0.47203226372084867, + "grad_norm": 46.21573257446289, + "learning_rate": 4.939535448220667e-05, + "loss": 0.7498, + "step": 673 + }, + { + "epoch": 0.47273364895668946, + "grad_norm": 12.326519966125488, + "learning_rate": 4.939218092537365e-05, + "loss": 0.1734, + "step": 674 + }, + { + "epoch": 0.47343503419253025, + "grad_norm": 2.402872085571289, + "learning_rate": 4.938899916440759e-05, + "loss": 0.2545, + "step": 675 + }, + { + "epoch": 0.47413641942837104, + "grad_norm": 4.058804512023926, + "learning_rate": 4.938580920037864e-05, + "loss": 0.2573, + "step": 676 + }, + { + "epoch": 0.4748378046642118, + "grad_norm": 11.151541709899902, + "learning_rate": 4.938261103435973e-05, + "loss": 0.4285, + "step": 677 + }, + { + "epoch": 0.4755391899000526, + "grad_norm": 6.957059860229492, + "learning_rate": 4.937940466742653e-05, + "loss": 0.2347, + "step": 678 + }, + { + "epoch": 0.4762405751358934, + "grad_norm": 5.085255146026611, + "learning_rate": 4.9376190100657495e-05, + "loss": 0.2671, + "step": 679 + }, + { + "epoch": 0.4769419603717342, + "grad_norm": 9.267155647277832, + "learning_rate": 4.93729673351338e-05, + "loss": 0.1125, + "step": 680 + }, + { + "epoch": 0.477643345607575, + "grad_norm": 6.162266731262207, + "learning_rate": 4.9369736371939426e-05, + "loss": 0.0683, + "step": 681 + }, + { + "epoch": 0.4783447308434157, + "grad_norm": 3.6411261558532715, + "learning_rate": 4.936649721216105e-05, + "loss": 0.0356, + "step": 682 + }, + { + "epoch": 0.4790461160792565, + "grad_norm": 48.720794677734375, + "learning_rate": 4.9363249856888165e-05, + "loss": 1.0375, + "step": 683 + }, + { + "epoch": 0.4797475013150973, + "grad_norm": 0.3436586856842041, + "learning_rate": 4.9359994307212994e-05, + "loss": 0.0027, + "step": 684 + }, + { + "epoch": 0.4804488865509381, + "grad_norm": 52.97607421875, + "learning_rate": 4.93567305642305e-05, + "loss": 1.6276, + "step": 685 + }, + { + "epoch": 0.4811502717867789, + "grad_norm": 0.09510497748851776, + "learning_rate": 4.935345862903844e-05, + "loss": 0.0007, + "step": 686 + }, + { + "epoch": 0.4818516570226197, + "grad_norm": 56.231292724609375, + "learning_rate": 4.93501785027373e-05, + "loss": 2.0342, + "step": 687 + }, + { + "epoch": 0.48255304225846046, + "grad_norm": 0.21096964180469513, + "learning_rate": 4.934689018643032e-05, + "loss": 0.0015, + "step": 688 + }, + { + "epoch": 0.48325442749430125, + "grad_norm": 0.19893679022789001, + "learning_rate": 4.93435936812235e-05, + "loss": 0.0014, + "step": 689 + }, + { + "epoch": 0.48395581273014204, + "grad_norm": 51.99702072143555, + "learning_rate": 4.9340288988225594e-05, + "loss": 1.0089, + "step": 690 + }, + { + "epoch": 0.48465719796598283, + "grad_norm": 53.519405364990234, + "learning_rate": 4.9336976108548126e-05, + "loss": 1.7908, + "step": 691 + }, + { + "epoch": 0.4853585832018236, + "grad_norm": 106.84630584716797, + "learning_rate": 4.9333655043305354e-05, + "loss": 3.6845, + "step": 692 + }, + { + "epoch": 0.4860599684376644, + "grad_norm": 0.24259038269519806, + "learning_rate": 4.9330325793614285e-05, + "loss": 0.0019, + "step": 693 + }, + { + "epoch": 0.48676135367350515, + "grad_norm": 0.8419340252876282, + "learning_rate": 4.93269883605947e-05, + "loss": 0.0072, + "step": 694 + }, + { + "epoch": 0.48746273890934594, + "grad_norm": 1.204994559288025, + "learning_rate": 4.932364274536912e-05, + "loss": 0.0087, + "step": 695 + }, + { + "epoch": 0.48816412414518673, + "grad_norm": 1.1827445030212402, + "learning_rate": 4.932028894906279e-05, + "loss": 0.0104, + "step": 696 + }, + { + "epoch": 0.4888655093810275, + "grad_norm": 0.5520339608192444, + "learning_rate": 4.931692697280378e-05, + "loss": 0.0045, + "step": 697 + }, + { + "epoch": 0.4895668946168683, + "grad_norm": 0.9636574983596802, + "learning_rate": 4.931355681772283e-05, + "loss": 0.0079, + "step": 698 + }, + { + "epoch": 0.4902682798527091, + "grad_norm": 0.5019407868385315, + "learning_rate": 4.931017848495348e-05, + "loss": 0.004, + "step": 699 + }, + { + "epoch": 0.4909696650885499, + "grad_norm": 0.0757029727101326, + "learning_rate": 4.9306791975632026e-05, + "loss": 0.0006, + "step": 700 + }, + { + "epoch": 0.4916710503243907, + "grad_norm": 110.13203430175781, + "learning_rate": 4.9303397290897465e-05, + "loss": 4.3731, + "step": 701 + }, + { + "epoch": 0.49237243556023147, + "grad_norm": 0.03152499720454216, + "learning_rate": 4.929999443189159e-05, + "loss": 0.0002, + "step": 702 + }, + { + "epoch": 0.49307382079607226, + "grad_norm": 0.029290443286299706, + "learning_rate": 4.929658339975893e-05, + "loss": 0.0002, + "step": 703 + }, + { + "epoch": 0.49377520603191305, + "grad_norm": 0.015484875068068504, + "learning_rate": 4.9293164195646754e-05, + "loss": 0.0001, + "step": 704 + }, + { + "epoch": 0.49447659126775384, + "grad_norm": 53.61540222167969, + "learning_rate": 4.928973682070509e-05, + "loss": 2.7716, + "step": 705 + }, + { + "epoch": 0.4951779765035946, + "grad_norm": 55.3106575012207, + "learning_rate": 4.9286301276086715e-05, + "loss": 2.3966, + "step": 706 + }, + { + "epoch": 0.49587936173943536, + "grad_norm": 0.014001145958900452, + "learning_rate": 4.928285756294714e-05, + "loss": 0.0001, + "step": 707 + }, + { + "epoch": 0.49658074697527615, + "grad_norm": 0.026434065774083138, + "learning_rate": 4.927940568244465e-05, + "loss": 0.0002, + "step": 708 + }, + { + "epoch": 0.49728213221111695, + "grad_norm": 0.04528193548321724, + "learning_rate": 4.927594563574023e-05, + "loss": 0.0003, + "step": 709 + }, + { + "epoch": 0.49798351744695774, + "grad_norm": 0.0950818806886673, + "learning_rate": 4.9272477423997666e-05, + "loss": 0.0007, + "step": 710 + }, + { + "epoch": 0.4986849026827985, + "grad_norm": 0.1095154732465744, + "learning_rate": 4.926900104838346e-05, + "loss": 0.0008, + "step": 711 + }, + { + "epoch": 0.4993862879186393, + "grad_norm": 0.13107261061668396, + "learning_rate": 4.9265516510066864e-05, + "loss": 0.001, + "step": 712 + }, + { + "epoch": 0.5000876731544801, + "grad_norm": 53.58122634887695, + "learning_rate": 4.926202381021988e-05, + "loss": 1.8605, + "step": 713 + }, + { + "epoch": 0.5007890583903208, + "grad_norm": 48.86884689331055, + "learning_rate": 4.925852295001725e-05, + "loss": 1.3723, + "step": 714 + }, + { + "epoch": 0.5014904436261617, + "grad_norm": 0.40202492475509644, + "learning_rate": 4.9255013930636465e-05, + "loss": 0.0033, + "step": 715 + }, + { + "epoch": 0.5021918288620024, + "grad_norm": 43.420196533203125, + "learning_rate": 4.925149675325775e-05, + "loss": 0.9394, + "step": 716 + }, + { + "epoch": 0.5028932140978433, + "grad_norm": 56.66259002685547, + "learning_rate": 4.9247971419064097e-05, + "loss": 1.2755, + "step": 717 + }, + { + "epoch": 0.503594599333684, + "grad_norm": 3.452730894088745, + "learning_rate": 4.9244437929241215e-05, + "loss": 0.0353, + "step": 718 + }, + { + "epoch": 0.5042959845695248, + "grad_norm": 4.454288005828857, + "learning_rate": 4.924089628497757e-05, + "loss": 0.0486, + "step": 719 + }, + { + "epoch": 0.5049973698053656, + "grad_norm": 18.87127685546875, + "learning_rate": 4.923734648746437e-05, + "loss": 0.3667, + "step": 720 + }, + { + "epoch": 0.5056987550412064, + "grad_norm": 11.85842514038086, + "learning_rate": 4.923378853789556e-05, + "loss": 0.3075, + "step": 721 + }, + { + "epoch": 0.5064001402770472, + "grad_norm": 9.63955307006836, + "learning_rate": 4.923022243746783e-05, + "loss": 0.1304, + "step": 722 + }, + { + "epoch": 0.507101525512888, + "grad_norm": 9.134459495544434, + "learning_rate": 4.922664818738063e-05, + "loss": 0.1346, + "step": 723 + }, + { + "epoch": 0.5078029107487287, + "grad_norm": 2.0840797424316406, + "learning_rate": 4.92230657888361e-05, + "loss": 0.2203, + "step": 724 + }, + { + "epoch": 0.5085042959845695, + "grad_norm": 8.857693672180176, + "learning_rate": 4.921947524303919e-05, + "loss": 0.1232, + "step": 725 + }, + { + "epoch": 0.5092056812204103, + "grad_norm": 8.82702350616455, + "learning_rate": 4.921587655119753e-05, + "loss": 0.1143, + "step": 726 + }, + { + "epoch": 0.5099070664562511, + "grad_norm": 4.017044544219971, + "learning_rate": 4.9212269714521505e-05, + "loss": 0.0433, + "step": 727 + }, + { + "epoch": 0.5106084516920919, + "grad_norm": 1.7636412382125854, + "learning_rate": 4.920865473422428e-05, + "loss": 0.0171, + "step": 728 + }, + { + "epoch": 0.5113098369279326, + "grad_norm": 107.01025390625, + "learning_rate": 4.920503161152171e-05, + "loss": 2.4773, + "step": 729 + }, + { + "epoch": 0.5120112221637735, + "grad_norm": 0.35363465547561646, + "learning_rate": 4.9201400347632404e-05, + "loss": 0.0028, + "step": 730 + }, + { + "epoch": 0.5127126073996142, + "grad_norm": 0.19405458867549896, + "learning_rate": 4.919776094377771e-05, + "loss": 0.0015, + "step": 731 + }, + { + "epoch": 0.5134139926354551, + "grad_norm": 55.00040054321289, + "learning_rate": 4.919411340118173e-05, + "loss": 1.9537, + "step": 732 + }, + { + "epoch": 0.5141153778712958, + "grad_norm": 55.760963439941406, + "learning_rate": 4.919045772107127e-05, + "loss": 2.1316, + "step": 733 + }, + { + "epoch": 0.5148167631071366, + "grad_norm": 55.24003601074219, + "learning_rate": 4.91867939046759e-05, + "loss": 2.1231, + "step": 734 + }, + { + "epoch": 0.5155181483429774, + "grad_norm": 0.03399452567100525, + "learning_rate": 4.918312195322792e-05, + "loss": 0.0002, + "step": 735 + }, + { + "epoch": 0.5162195335788181, + "grad_norm": 204.8887176513672, + "learning_rate": 4.917944186796235e-05, + "loss": 8.6129, + "step": 736 + }, + { + "epoch": 0.516920918814659, + "grad_norm": 124.1363754272461, + "learning_rate": 4.917575365011699e-05, + "loss": 3.9954, + "step": 737 + }, + { + "epoch": 0.5176223040504997, + "grad_norm": 0.14609888195991516, + "learning_rate": 4.9172057300932306e-05, + "loss": 0.0011, + "step": 738 + }, + { + "epoch": 0.5183236892863405, + "grad_norm": 0.33696234226226807, + "learning_rate": 4.9168352821651575e-05, + "loss": 0.0027, + "step": 739 + }, + { + "epoch": 0.5190250745221813, + "grad_norm": 105.28889465332031, + "learning_rate": 4.916464021352075e-05, + "loss": 3.0367, + "step": 740 + }, + { + "epoch": 0.5197264597580221, + "grad_norm": 1.017066240310669, + "learning_rate": 4.916091947778855e-05, + "loss": 0.0088, + "step": 741 + }, + { + "epoch": 0.5204278449938629, + "grad_norm": 1.143968939781189, + "learning_rate": 4.91571906157064e-05, + "loss": 0.0102, + "step": 742 + }, + { + "epoch": 0.5211292302297037, + "grad_norm": 0.7457045912742615, + "learning_rate": 4.91534536285285e-05, + "loss": 0.0063, + "step": 743 + }, + { + "epoch": 0.5218306154655444, + "grad_norm": 0.700978696346283, + "learning_rate": 4.9149708517511743e-05, + "loss": 0.0058, + "step": 744 + }, + { + "epoch": 0.5225320007013853, + "grad_norm": 0.38463106751441956, + "learning_rate": 4.914595528391578e-05, + "loss": 0.0031, + "step": 745 + }, + { + "epoch": 0.523233385937226, + "grad_norm": 0.14777745306491852, + "learning_rate": 4.914219392900297e-05, + "loss": 0.0011, + "step": 746 + }, + { + "epoch": 0.5239347711730669, + "grad_norm": 0.07884679734706879, + "learning_rate": 4.913842445403843e-05, + "loss": 0.0006, + "step": 747 + }, + { + "epoch": 0.5246361564089076, + "grad_norm": 61.55155563354492, + "learning_rate": 4.913464686028999e-05, + "loss": 1.992, + "step": 748 + }, + { + "epoch": 0.5253375416447483, + "grad_norm": 0.027440281584858894, + "learning_rate": 4.913086114902822e-05, + "loss": 0.0002, + "step": 749 + }, + { + "epoch": 0.5260389268805892, + "grad_norm": 55.66230010986328, + "learning_rate": 4.912706732152641e-05, + "loss": 2.231, + "step": 750 + }, + { + "epoch": 0.5267403121164299, + "grad_norm": 54.453102111816406, + "learning_rate": 4.912326537906059e-05, + "loss": 2.0983, + "step": 751 + }, + { + "epoch": 0.5274416973522708, + "grad_norm": 0.0637647733092308, + "learning_rate": 4.911945532290951e-05, + "loss": 0.0005, + "step": 752 + }, + { + "epoch": 0.5281430825881115, + "grad_norm": 0.13056063652038574, + "learning_rate": 4.9115637154354674e-05, + "loss": 0.001, + "step": 753 + }, + { + "epoch": 0.5288444678239523, + "grad_norm": 52.72209167480469, + "learning_rate": 4.911181087468027e-05, + "loss": 1.5666, + "step": 754 + }, + { + "epoch": 0.5295458530597931, + "grad_norm": 48.385467529296875, + "learning_rate": 4.9107976485173255e-05, + "loss": 1.2941, + "step": 755 + }, + { + "epoch": 0.5302472382956339, + "grad_norm": 0.5080452561378479, + "learning_rate": 4.91041339871233e-05, + "loss": 0.0041, + "step": 756 + }, + { + "epoch": 0.5309486235314747, + "grad_norm": 80.44468688964844, + "learning_rate": 4.910028338182279e-05, + "loss": 1.6558, + "step": 757 + }, + { + "epoch": 0.5316500087673155, + "grad_norm": 1.59328031539917, + "learning_rate": 4.909642467056685e-05, + "loss": 0.0141, + "step": 758 + }, + { + "epoch": 0.5323513940031562, + "grad_norm": 2.678886651992798, + "learning_rate": 4.9092557854653344e-05, + "loss": 0.0255, + "step": 759 + }, + { + "epoch": 0.533052779238997, + "grad_norm": 2.742159366607666, + "learning_rate": 4.908868293538282e-05, + "loss": 0.0271, + "step": 760 + }, + { + "epoch": 0.5337541644748378, + "grad_norm": 2.275885820388794, + "learning_rate": 4.908479991405861e-05, + "loss": 0.0215, + "step": 761 + }, + { + "epoch": 0.5344555497106785, + "grad_norm": 1.4059484004974365, + "learning_rate": 4.9080908791986724e-05, + "loss": 0.0126, + "step": 762 + }, + { + "epoch": 0.5351569349465194, + "grad_norm": 0.7012129426002502, + "learning_rate": 4.907700957047592e-05, + "loss": 0.0058, + "step": 763 + }, + { + "epoch": 0.5358583201823601, + "grad_norm": 0.2713419497013092, + "learning_rate": 4.9073102250837655e-05, + "loss": 0.0021, + "step": 764 + }, + { + "epoch": 0.536559705418201, + "grad_norm": 0.13228341937065125, + "learning_rate": 4.9069186834386145e-05, + "loss": 0.001, + "step": 765 + }, + { + "epoch": 0.5372610906540417, + "grad_norm": 0.06449166685342789, + "learning_rate": 4.90652633224383e-05, + "loss": 0.0005, + "step": 766 + }, + { + "epoch": 0.5379624758898826, + "grad_norm": 0.017581632360816002, + "learning_rate": 4.9061331716313783e-05, + "loss": 0.0001, + "step": 767 + }, + { + "epoch": 0.5386638611257233, + "grad_norm": 0.016885504126548767, + "learning_rate": 4.905739201733495e-05, + "loss": 0.0001, + "step": 768 + }, + { + "epoch": 0.5393652463615641, + "grad_norm": 0.005497175268828869, + "learning_rate": 4.9053444226826886e-05, + "loss": 0.0, + "step": 769 + }, + { + "epoch": 0.5400666315974049, + "grad_norm": 58.73657989501953, + "learning_rate": 4.90494883461174e-05, + "loss": 3.1923, + "step": 770 + }, + { + "epoch": 0.5407680168332457, + "grad_norm": 0.0022041036281734705, + "learning_rate": 4.9045524376537034e-05, + "loss": 0.0, + "step": 771 + }, + { + "epoch": 0.5414694020690864, + "grad_norm": 52.654136657714844, + "learning_rate": 4.9041552319419035e-05, + "loss": 3.097, + "step": 772 + }, + { + "epoch": 0.5421707873049272, + "grad_norm": 0.0026611709035933018, + "learning_rate": 4.903757217609938e-05, + "loss": 0.0, + "step": 773 + }, + { + "epoch": 0.542872172540768, + "grad_norm": 0.0028983913362026215, + "learning_rate": 4.9033583947916765e-05, + "loss": 0.0, + "step": 774 + }, + { + "epoch": 0.5435735577766088, + "grad_norm": 105.84739685058594, + "learning_rate": 4.902958763621258e-05, + "loss": 5.9716, + "step": 775 + }, + { + "epoch": 0.5442749430124496, + "grad_norm": 54.44437789916992, + "learning_rate": 4.902558324233098e-05, + "loss": 2.7511, + "step": 776 + }, + { + "epoch": 0.5449763282482903, + "grad_norm": 0.009829580783843994, + "learning_rate": 4.9021570767618794e-05, + "loss": 0.0001, + "step": 777 + }, + { + "epoch": 0.5456777134841312, + "grad_norm": 53.416908264160156, + "learning_rate": 4.90175502134256e-05, + "loss": 2.337, + "step": 778 + }, + { + "epoch": 0.5463790987199719, + "grad_norm": 0.050978660583496094, + "learning_rate": 4.9013521581103686e-05, + "loss": 0.0004, + "step": 779 + }, + { + "epoch": 0.5470804839558128, + "grad_norm": 0.11917990446090698, + "learning_rate": 4.900948487200804e-05, + "loss": 0.0009, + "step": 780 + }, + { + "epoch": 0.5477818691916535, + "grad_norm": 48.541015625, + "learning_rate": 4.900544008749639e-05, + "loss": 1.2937, + "step": 781 + }, + { + "epoch": 0.5484832544274943, + "grad_norm": 0.3613051474094391, + "learning_rate": 4.900138722892917e-05, + "loss": 0.003, + "step": 782 + }, + { + "epoch": 0.5491846396633351, + "grad_norm": 35.976261138916016, + "learning_rate": 4.899732629766953e-05, + "loss": 0.7424, + "step": 783 + }, + { + "epoch": 0.5498860248991758, + "grad_norm": 1.3365657329559326, + "learning_rate": 4.899325729508333e-05, + "loss": 0.012, + "step": 784 + }, + { + "epoch": 0.5505874101350167, + "grad_norm": 2.2534306049346924, + "learning_rate": 4.898918022253916e-05, + "loss": 0.0228, + "step": 785 + }, + { + "epoch": 0.5512887953708574, + "grad_norm": 24.77885627746582, + "learning_rate": 4.898509508140829e-05, + "loss": 0.4601, + "step": 786 + }, + { + "epoch": 0.5519901806066982, + "grad_norm": 4.28820276260376, + "learning_rate": 4.898100187306475e-05, + "loss": 0.0482, + "step": 787 + }, + { + "epoch": 0.552691565842539, + "grad_norm": 4.493249416351318, + "learning_rate": 4.897690059888524e-05, + "loss": 0.0503, + "step": 788 + }, + { + "epoch": 0.5533929510783798, + "grad_norm": 3.269726037979126, + "learning_rate": 4.897279126024923e-05, + "loss": 0.0349, + "step": 789 + }, + { + "epoch": 0.5540943363142206, + "grad_norm": 3.0403542518615723, + "learning_rate": 4.896867385853883e-05, + "loss": 0.0316, + "step": 790 + }, + { + "epoch": 0.5547957215500614, + "grad_norm": 25.840961456298828, + "learning_rate": 4.896454839513892e-05, + "loss": 0.4405, + "step": 791 + }, + { + "epoch": 0.5554971067859021, + "grad_norm": 72.00930786132812, + "learning_rate": 4.8960414871437055e-05, + "loss": 1.4578, + "step": 792 + }, + { + "epoch": 0.556198492021743, + "grad_norm": 0.7836804389953613, + "learning_rate": 4.895627328882352e-05, + "loss": 0.0068, + "step": 793 + }, + { + "epoch": 0.5568998772575837, + "grad_norm": 0.7037481069564819, + "learning_rate": 4.895212364869132e-05, + "loss": 0.0061, + "step": 794 + }, + { + "epoch": 0.5576012624934246, + "grad_norm": 0.4775547981262207, + "learning_rate": 4.894796595243614e-05, + "loss": 0.004, + "step": 795 + }, + { + "epoch": 0.5583026477292653, + "grad_norm": 0.4537871778011322, + "learning_rate": 4.894380020145639e-05, + "loss": 0.0038, + "step": 796 + }, + { + "epoch": 0.559004032965106, + "grad_norm": 0.2530267536640167, + "learning_rate": 4.8939626397153196e-05, + "loss": 0.0021, + "step": 797 + }, + { + "epoch": 0.5597054182009469, + "grad_norm": 0.08755356073379517, + "learning_rate": 4.8935444540930385e-05, + "loss": 0.0007, + "step": 798 + }, + { + "epoch": 0.5604068034367876, + "grad_norm": 0.058386508375406265, + "learning_rate": 4.8931254634194504e-05, + "loss": 0.0004, + "step": 799 + }, + { + "epoch": 0.5611081886726285, + "grad_norm": 0.020555956289172173, + "learning_rate": 4.892705667835479e-05, + "loss": 0.0002, + "step": 800 + }, + { + "epoch": 0.5618095739084692, + "grad_norm": 0.023624928668141365, + "learning_rate": 4.892285067482319e-05, + "loss": 0.0002, + "step": 801 + }, + { + "epoch": 0.56251095914431, + "grad_norm": 54.471675872802734, + "learning_rate": 4.891863662501437e-05, + "loss": 2.8657, + "step": 802 + }, + { + "epoch": 0.5632123443801508, + "grad_norm": 0.005444019101560116, + "learning_rate": 4.8914414530345684e-05, + "loss": 0.0, + "step": 803 + }, + { + "epoch": 0.5639137296159916, + "grad_norm": 0.008992165327072144, + "learning_rate": 4.891018439223722e-05, + "loss": 0.0001, + "step": 804 + }, + { + "epoch": 0.5646151148518324, + "grad_norm": 0.0062455967999994755, + "learning_rate": 4.890594621211174e-05, + "loss": 0.0, + "step": 805 + }, + { + "epoch": 0.5653165000876732, + "grad_norm": 0.006425914354622364, + "learning_rate": 4.8901699991394735e-05, + "loss": 0.0, + "step": 806 + }, + { + "epoch": 0.5660178853235139, + "grad_norm": 0.008212440647184849, + "learning_rate": 4.8897445731514384e-05, + "loss": 0.0001, + "step": 807 + }, + { + "epoch": 0.5667192705593547, + "grad_norm": 0.004950184375047684, + "learning_rate": 4.889318343390158e-05, + "loss": 0.0, + "step": 808 + }, + { + "epoch": 0.5674206557951955, + "grad_norm": 0.0056209079921245575, + "learning_rate": 4.8888913099989916e-05, + "loss": 0.0, + "step": 809 + }, + { + "epoch": 0.5681220410310363, + "grad_norm": 0.0028246166184544563, + "learning_rate": 4.888463473121568e-05, + "loss": 0.0, + "step": 810 + }, + { + "epoch": 0.5688234262668771, + "grad_norm": 53.988868713378906, + "learning_rate": 4.888034832901788e-05, + "loss": 2.7359, + "step": 811 + }, + { + "epoch": 0.5695248115027178, + "grad_norm": 0.004621135536581278, + "learning_rate": 4.8876053894838215e-05, + "loss": 0.0, + "step": 812 + }, + { + "epoch": 0.5702261967385587, + "grad_norm": 0.011877781711518764, + "learning_rate": 4.887175143012108e-05, + "loss": 0.0001, + "step": 813 + }, + { + "epoch": 0.5709275819743994, + "grad_norm": 65.13773345947266, + "learning_rate": 4.886744093631358e-05, + "loss": 3.5609, + "step": 814 + }, + { + "epoch": 0.5716289672102403, + "grad_norm": 161.86300659179688, + "learning_rate": 4.886312241486554e-05, + "loss": 7.2583, + "step": 815 + }, + { + "epoch": 0.572330352446081, + "grad_norm": 0.032078277319669724, + "learning_rate": 4.8858795867229425e-05, + "loss": 0.0002, + "step": 816 + }, + { + "epoch": 0.5730317376819218, + "grad_norm": 113.32162475585938, + "learning_rate": 4.8854461294860454e-05, + "loss": 4.1658, + "step": 817 + }, + { + "epoch": 0.5737331229177626, + "grad_norm": 0.17656919360160828, + "learning_rate": 4.885011869921654e-05, + "loss": 0.0014, + "step": 818 + }, + { + "epoch": 0.5744345081536034, + "grad_norm": 0.2188340574502945, + "learning_rate": 4.884576808175828e-05, + "loss": 0.0018, + "step": 819 + }, + { + "epoch": 0.5751358933894442, + "grad_norm": 0.38712555170059204, + "learning_rate": 4.884140944394896e-05, + "loss": 0.0033, + "step": 820 + }, + { + "epoch": 0.5758372786252849, + "grad_norm": 0.7210093140602112, + "learning_rate": 4.883704278725458e-05, + "loss": 0.0062, + "step": 821 + }, + { + "epoch": 0.5765386638611257, + "grad_norm": 0.7694672346115112, + "learning_rate": 4.883266811314385e-05, + "loss": 0.0069, + "step": 822 + }, + { + "epoch": 0.5772400490969665, + "grad_norm": 0.6353862881660461, + "learning_rate": 4.8828285423088146e-05, + "loss": 0.0056, + "step": 823 + }, + { + "epoch": 0.5779414343328073, + "grad_norm": 87.86915588378906, + "learning_rate": 4.8823894718561545e-05, + "loss": 2.1381, + "step": 824 + }, + { + "epoch": 0.578642819568648, + "grad_norm": 0.42429307103157043, + "learning_rate": 4.8819496001040846e-05, + "loss": 0.0036, + "step": 825 + }, + { + "epoch": 0.5793442048044889, + "grad_norm": 0.40686339139938354, + "learning_rate": 4.881508927200552e-05, + "loss": 0.0034, + "step": 826 + }, + { + "epoch": 0.5800455900403296, + "grad_norm": 0.27934542298316956, + "learning_rate": 4.881067453293774e-05, + "loss": 0.0022, + "step": 827 + }, + { + "epoch": 0.5807469752761705, + "grad_norm": 44.31092834472656, + "learning_rate": 4.8806251785322355e-05, + "loss": 1.1003, + "step": 828 + }, + { + "epoch": 0.5814483605120112, + "grad_norm": 0.17587609589099884, + "learning_rate": 4.880182103064695e-05, + "loss": 0.0014, + "step": 829 + }, + { + "epoch": 0.582149745747852, + "grad_norm": 0.2651273012161255, + "learning_rate": 4.879738227040176e-05, + "loss": 0.0021, + "step": 830 + }, + { + "epoch": 0.5828511309836928, + "grad_norm": 149.13035583496094, + "learning_rate": 4.879293550607974e-05, + "loss": 2.9684, + "step": 831 + }, + { + "epoch": 0.5835525162195335, + "grad_norm": 0.17069180309772491, + "learning_rate": 4.878848073917651e-05, + "loss": 0.0014, + "step": 832 + }, + { + "epoch": 0.5842539014553744, + "grad_norm": 0.24799254536628723, + "learning_rate": 4.878401797119042e-05, + "loss": 0.002, + "step": 833 + }, + { + "epoch": 0.5849552866912151, + "grad_norm": 0.15171018242835999, + "learning_rate": 4.8779547203622477e-05, + "loss": 0.0012, + "step": 834 + }, + { + "epoch": 0.585656671927056, + "grad_norm": 45.8453254699707, + "learning_rate": 4.877506843797639e-05, + "loss": 1.2116, + "step": 835 + }, + { + "epoch": 0.5863580571628967, + "grad_norm": 70.00684356689453, + "learning_rate": 4.877058167575857e-05, + "loss": 2.1496, + "step": 836 + }, + { + "epoch": 0.5870594423987375, + "grad_norm": 43.86625671386719, + "learning_rate": 4.876608691847809e-05, + "loss": 1.077, + "step": 837 + }, + { + "epoch": 0.5877608276345783, + "grad_norm": 39.75180435180664, + "learning_rate": 4.876158416764675e-05, + "loss": 0.8959, + "step": 838 + }, + { + "epoch": 0.5884622128704191, + "grad_norm": 78.85111999511719, + "learning_rate": 4.8757073424779e-05, + "loss": 1.609, + "step": 839 + }, + { + "epoch": 0.5891635981062598, + "grad_norm": 2.562579393386841, + "learning_rate": 4.8752554691392004e-05, + "loss": 0.0262, + "step": 840 + }, + { + "epoch": 0.5898649833421007, + "grad_norm": 3.174607276916504, + "learning_rate": 4.8748027969005606e-05, + "loss": 0.0331, + "step": 841 + }, + { + "epoch": 0.5905663685779414, + "grad_norm": 24.42951202392578, + "learning_rate": 4.874349325914232e-05, + "loss": 0.494, + "step": 842 + }, + { + "epoch": 0.5912677538137823, + "grad_norm": 5.76008415222168, + "learning_rate": 4.87389505633274e-05, + "loss": 0.069, + "step": 843 + }, + { + "epoch": 0.591969139049623, + "grad_norm": 16.484464645385742, + "learning_rate": 4.873439988308871e-05, + "loss": 0.3677, + "step": 844 + }, + { + "epoch": 0.5926705242854637, + "grad_norm": 8.441133499145508, + "learning_rate": 4.872984121995686e-05, + "loss": 0.1122, + "step": 845 + }, + { + "epoch": 0.5933719095213046, + "grad_norm": 7.171937942504883, + "learning_rate": 4.8725274575465106e-05, + "loss": 0.093, + "step": 846 + }, + { + "epoch": 0.5940732947571453, + "grad_norm": 7.630092620849609, + "learning_rate": 4.872069995114943e-05, + "loss": 0.0997, + "step": 847 + }, + { + "epoch": 0.5947746799929862, + "grad_norm": 5.521340370178223, + "learning_rate": 4.871611734854845e-05, + "loss": 0.065, + "step": 848 + }, + { + "epoch": 0.5954760652288269, + "grad_norm": 40.95110321044922, + "learning_rate": 4.871152676920351e-05, + "loss": 0.6631, + "step": 849 + }, + { + "epoch": 0.5961774504646677, + "grad_norm": 2.602055311203003, + "learning_rate": 4.8706928214658596e-05, + "loss": 0.0269, + "step": 850 + }, + { + "epoch": 0.5968788357005085, + "grad_norm": 30.752164840698242, + "learning_rate": 4.8702321686460424e-05, + "loss": 0.5944, + "step": 851 + }, + { + "epoch": 0.5975802209363493, + "grad_norm": 34.97576904296875, + "learning_rate": 4.8697707186158345e-05, + "loss": 0.7216, + "step": 852 + }, + { + "epoch": 0.5982816061721901, + "grad_norm": 1.3377373218536377, + "learning_rate": 4.869308471530443e-05, + "loss": 0.0126, + "step": 853 + }, + { + "epoch": 0.5989829914080309, + "grad_norm": 33.73777389526367, + "learning_rate": 4.868845427545341e-05, + "loss": 0.6655, + "step": 854 + }, + { + "epoch": 0.5996843766438716, + "grad_norm": 27.722604751586914, + "learning_rate": 4.868381586816268e-05, + "loss": 0.5284, + "step": 855 + }, + { + "epoch": 0.6003857618797125, + "grad_norm": 39.0617561340332, + "learning_rate": 4.867916949499236e-05, + "loss": 0.9141, + "step": 856 + }, + { + "epoch": 0.6010871471155532, + "grad_norm": 25.998207092285156, + "learning_rate": 4.8674515157505205e-05, + "loss": 0.4843, + "step": 857 + }, + { + "epoch": 0.601788532351394, + "grad_norm": 70.33052062988281, + "learning_rate": 4.8669852857266686e-05, + "loss": 1.1847, + "step": 858 + }, + { + "epoch": 0.6024899175872348, + "grad_norm": 2.859663724899292, + "learning_rate": 4.866518259584492e-05, + "loss": 0.1724, + "step": 859 + }, + { + "epoch": 0.6031913028230755, + "grad_norm": 7.6815185546875, + "learning_rate": 4.866050437481072e-05, + "loss": 0.2053, + "step": 860 + }, + { + "epoch": 0.6038926880589164, + "grad_norm": 15.574623107910156, + "learning_rate": 4.8655818195737576e-05, + "loss": 0.2786, + "step": 861 + }, + { + "epoch": 0.6045940732947571, + "grad_norm": 17.501440048217773, + "learning_rate": 4.865112406020164e-05, + "loss": 0.3404, + "step": 862 + }, + { + "epoch": 0.605295458530598, + "grad_norm": 18.48247718811035, + "learning_rate": 4.8646421969781757e-05, + "loss": 0.3676, + "step": 863 + }, + { + "epoch": 0.6059968437664387, + "grad_norm": 15.630892753601074, + "learning_rate": 4.864171192605944e-05, + "loss": 0.2856, + "step": 864 + }, + { + "epoch": 0.6066982290022795, + "grad_norm": 15.12441349029541, + "learning_rate": 4.863699393061888e-05, + "loss": 0.2728, + "step": 865 + }, + { + "epoch": 0.6073996142381203, + "grad_norm": 10.427019119262695, + "learning_rate": 4.863226798504694e-05, + "loss": 0.308, + "step": 866 + }, + { + "epoch": 0.6081009994739611, + "grad_norm": 10.065474510192871, + "learning_rate": 4.8627534090933156e-05, + "loss": 0.1617, + "step": 867 + }, + { + "epoch": 0.6088023847098019, + "grad_norm": 11.481346130371094, + "learning_rate": 4.862279224986973e-05, + "loss": 0.2136, + "step": 868 + }, + { + "epoch": 0.6095037699456426, + "grad_norm": 5.78438663482666, + "learning_rate": 4.8618042463451566e-05, + "loss": 0.0726, + "step": 869 + }, + { + "epoch": 0.6102051551814834, + "grad_norm": 39.987937927246094, + "learning_rate": 4.8613284733276206e-05, + "loss": 0.8901, + "step": 870 + }, + { + "epoch": 0.6109065404173242, + "grad_norm": 3.24749755859375, + "learning_rate": 4.860851906094388e-05, + "loss": 0.0369, + "step": 871 + }, + { + "epoch": 0.611607925653165, + "grad_norm": 2.4093921184539795, + "learning_rate": 4.860374544805749e-05, + "loss": 0.0254, + "step": 872 + }, + { + "epoch": 0.6123093108890058, + "grad_norm": 0.970287024974823, + "learning_rate": 4.85989638962226e-05, + "loss": 0.0092, + "step": 873 + }, + { + "epoch": 0.6130106961248466, + "grad_norm": 0.4123772084712982, + "learning_rate": 4.859417440704746e-05, + "loss": 0.0036, + "step": 874 + }, + { + "epoch": 0.6137120813606873, + "grad_norm": 0.14014503359794617, + "learning_rate": 4.8589376982142977e-05, + "loss": 0.0011, + "step": 875 + }, + { + "epoch": 0.6144134665965282, + "grad_norm": 0.05853430926799774, + "learning_rate": 4.858457162312272e-05, + "loss": 0.0005, + "step": 876 + }, + { + "epoch": 0.6151148518323689, + "grad_norm": 104.4763412475586, + "learning_rate": 4.857975833160295e-05, + "loss": 4.5228, + "step": 877 + }, + { + "epoch": 0.6158162370682098, + "grad_norm": 60.22207260131836, + "learning_rate": 4.857493710920258e-05, + "loss": 3.2127, + "step": 878 + }, + { + "epoch": 0.6165176223040505, + "grad_norm": 0.014606939628720284, + "learning_rate": 4.8570107957543185e-05, + "loss": 0.0001, + "step": 879 + }, + { + "epoch": 0.6172190075398913, + "grad_norm": 52.49748229980469, + "learning_rate": 4.856527087824902e-05, + "loss": 2.5181, + "step": 880 + }, + { + "epoch": 0.6179203927757321, + "grad_norm": 60.33863067626953, + "learning_rate": 4.8560425872947004e-05, + "loss": 3.4192, + "step": 881 + }, + { + "epoch": 0.6186217780115728, + "grad_norm": 0.017534319311380386, + "learning_rate": 4.8555572943266715e-05, + "loss": 0.0001, + "step": 882 + }, + { + "epoch": 0.6193231632474137, + "grad_norm": 0.03253171965479851, + "learning_rate": 4.85507120908404e-05, + "loss": 0.0002, + "step": 883 + }, + { + "epoch": 0.6200245484832544, + "grad_norm": 100.64522552490234, + "learning_rate": 4.8545843317302983e-05, + "loss": 3.6464, + "step": 884 + }, + { + "epoch": 0.6207259337190952, + "grad_norm": 52.276100158691406, + "learning_rate": 4.854096662429203e-05, + "loss": 1.5974, + "step": 885 + }, + { + "epoch": 0.621427318954936, + "grad_norm": 0.26515480875968933, + "learning_rate": 4.853608201344778e-05, + "loss": 0.0022, + "step": 886 + }, + { + "epoch": 0.6221287041907768, + "grad_norm": 0.4160686731338501, + "learning_rate": 4.853118948641314e-05, + "loss": 0.0036, + "step": 887 + }, + { + "epoch": 0.6228300894266176, + "grad_norm": 40.08952331542969, + "learning_rate": 4.852628904483367e-05, + "loss": 0.9215, + "step": 888 + }, + { + "epoch": 0.6235314746624584, + "grad_norm": 35.90172576904297, + "learning_rate": 4.852138069035761e-05, + "loss": 0.7604, + "step": 889 + }, + { + "epoch": 0.6242328598982991, + "grad_norm": 2.5494091510772705, + "learning_rate": 4.851646442463585e-05, + "loss": 0.0269, + "step": 890 + }, + { + "epoch": 0.62493424513414, + "grad_norm": 4.3015456199646, + "learning_rate": 4.851154024932193e-05, + "loss": 0.0495, + "step": 891 + }, + { + "epoch": 0.6256356303699807, + "grad_norm": 3.570831537246704, + "learning_rate": 4.8506608166072054e-05, + "loss": 0.0385, + "step": 892 + }, + { + "epoch": 0.6263370156058214, + "grad_norm": 3.575366735458374, + "learning_rate": 4.8501668176545115e-05, + "loss": 0.0397, + "step": 893 + }, + { + "epoch": 0.6270384008416623, + "grad_norm": 2.7419331073760986, + "learning_rate": 4.849672028240262e-05, + "loss": 0.0288, + "step": 894 + }, + { + "epoch": 0.627739786077503, + "grad_norm": 41.37546157836914, + "learning_rate": 4.849176448530878e-05, + "loss": 0.8202, + "step": 895 + }, + { + "epoch": 0.6284411713133439, + "grad_norm": 34.27451705932617, + "learning_rate": 4.8486800786930415e-05, + "loss": 0.7397, + "step": 896 + }, + { + "epoch": 0.6291425565491846, + "grad_norm": 1.0924946069717407, + "learning_rate": 4.848182918893705e-05, + "loss": 0.0104, + "step": 897 + }, + { + "epoch": 0.6298439417850255, + "grad_norm": 0.9149582386016846, + "learning_rate": 4.8476849693000836e-05, + "loss": 0.0085, + "step": 898 + }, + { + "epoch": 0.6305453270208662, + "grad_norm": 0.4877033829689026, + "learning_rate": 4.8471862300796594e-05, + "loss": 0.0043, + "step": 899 + }, + { + "epoch": 0.631246712256707, + "grad_norm": 50.69717788696289, + "learning_rate": 4.846686701400179e-05, + "loss": 1.3647, + "step": 900 + }, + { + "epoch": 0.6319480974925478, + "grad_norm": 45.04077911376953, + "learning_rate": 4.846186383429657e-05, + "loss": 1.366, + "step": 901 + }, + { + "epoch": 0.6326494827283886, + "grad_norm": 47.341819763183594, + "learning_rate": 4.8456852763363706e-05, + "loss": 1.2603, + "step": 902 + }, + { + "epoch": 0.6333508679642293, + "grad_norm": 44.485530853271484, + "learning_rate": 4.8451833802888625e-05, + "loss": 1.3248, + "step": 903 + }, + { + "epoch": 0.6340522532000702, + "grad_norm": 0.46667349338531494, + "learning_rate": 4.844680695455943e-05, + "loss": 0.0041, + "step": 904 + }, + { + "epoch": 0.6347536384359109, + "grad_norm": 0.7628422379493713, + "learning_rate": 4.844177222006687e-05, + "loss": 0.0067, + "step": 905 + }, + { + "epoch": 0.6354550236717517, + "grad_norm": 0.7451947927474976, + "learning_rate": 4.8436729601104327e-05, + "loss": 0.0069, + "step": 906 + }, + { + "epoch": 0.6361564089075925, + "grad_norm": 0.6644980907440186, + "learning_rate": 4.843167909936786e-05, + "loss": 0.006, + "step": 907 + }, + { + "epoch": 0.6368577941434332, + "grad_norm": 41.53312683105469, + "learning_rate": 4.8426620716556164e-05, + "loss": 1.1098, + "step": 908 + }, + { + "epoch": 0.6375591793792741, + "grad_norm": 0.5270825624465942, + "learning_rate": 4.8421554454370596e-05, + "loss": 0.0046, + "step": 909 + }, + { + "epoch": 0.6382605646151148, + "grad_norm": 90.92555236816406, + "learning_rate": 4.841648031451514e-05, + "loss": 2.5683, + "step": 910 + }, + { + "epoch": 0.6389619498509557, + "grad_norm": 46.597808837890625, + "learning_rate": 4.841139829869646e-05, + "loss": 1.4381, + "step": 911 + }, + { + "epoch": 0.6396633350867964, + "grad_norm": 0.9347837567329407, + "learning_rate": 4.8406308408623865e-05, + "loss": 0.0083, + "step": 912 + }, + { + "epoch": 0.6403647203226372, + "grad_norm": 0.8483829498291016, + "learning_rate": 4.840121064600928e-05, + "loss": 0.0077, + "step": 913 + }, + { + "epoch": 0.641066105558478, + "grad_norm": 37.393951416015625, + "learning_rate": 4.8396105012567315e-05, + "loss": 0.8573, + "step": 914 + }, + { + "epoch": 0.6417674907943188, + "grad_norm": 69.15557098388672, + "learning_rate": 4.839099151001521e-05, + "loss": 1.491, + "step": 915 + }, + { + "epoch": 0.6424688760301596, + "grad_norm": 1.3047503232955933, + "learning_rate": 4.838587014007284e-05, + "loss": 0.0126, + "step": 916 + }, + { + "epoch": 0.6431702612660003, + "grad_norm": 1.462365746498108, + "learning_rate": 4.838074090446276e-05, + "loss": 0.0144, + "step": 917 + }, + { + "epoch": 0.6438716465018411, + "grad_norm": 1.4955204725265503, + "learning_rate": 4.8375603804910154e-05, + "loss": 0.0146, + "step": 918 + }, + { + "epoch": 0.6445730317376819, + "grad_norm": 1.0593093633651733, + "learning_rate": 4.837045884314283e-05, + "loss": 0.0101, + "step": 919 + }, + { + "epoch": 0.6452744169735227, + "grad_norm": 0.5433834791183472, + "learning_rate": 4.836530602089127e-05, + "loss": 0.0049, + "step": 920 + }, + { + "epoch": 0.6459758022093635, + "grad_norm": 44.653175354003906, + "learning_rate": 4.836014533988857e-05, + "loss": 1.2462, + "step": 921 + }, + { + "epoch": 0.6466771874452043, + "grad_norm": 0.15624181926250458, + "learning_rate": 4.835497680187052e-05, + "loss": 0.0013, + "step": 922 + }, + { + "epoch": 0.647378572681045, + "grad_norm": 0.13333559036254883, + "learning_rate": 4.8349800408575506e-05, + "loss": 0.0011, + "step": 923 + }, + { + "epoch": 0.6480799579168859, + "grad_norm": 0.08697400987148285, + "learning_rate": 4.834461616174456e-05, + "loss": 0.0007, + "step": 924 + }, + { + "epoch": 0.6487813431527266, + "grad_norm": 49.62416076660156, + "learning_rate": 4.8339424063121375e-05, + "loss": 2.1264, + "step": 925 + }, + { + "epoch": 0.6494827283885675, + "grad_norm": 47.912906646728516, + "learning_rate": 4.833422411445228e-05, + "loss": 1.8048, + "step": 926 + }, + { + "epoch": 0.6501841136244082, + "grad_norm": 0.0816919356584549, + "learning_rate": 4.832901631748622e-05, + "loss": 0.0007, + "step": 927 + }, + { + "epoch": 0.650885498860249, + "grad_norm": 49.09192657470703, + "learning_rate": 4.8323800673974826e-05, + "loss": 1.7493, + "step": 928 + }, + { + "epoch": 0.6515868840960898, + "grad_norm": 98.90199279785156, + "learning_rate": 4.8318577185672325e-05, + "loss": 3.7461, + "step": 929 + }, + { + "epoch": 0.6522882693319305, + "grad_norm": 0.22995713353157043, + "learning_rate": 4.831334585433561e-05, + "loss": 0.002, + "step": 930 + }, + { + "epoch": 0.6529896545677714, + "grad_norm": 0.4566737413406372, + "learning_rate": 4.830810668172419e-05, + "loss": 0.0041, + "step": 931 + }, + { + "epoch": 0.6536910398036121, + "grad_norm": 0.5328819155693054, + "learning_rate": 4.830285966960023e-05, + "loss": 0.0048, + "step": 932 + }, + { + "epoch": 0.6543924250394529, + "grad_norm": 0.6611910462379456, + "learning_rate": 4.829760481972852e-05, + "loss": 0.0061, + "step": 933 + }, + { + "epoch": 0.6550938102752937, + "grad_norm": 76.24475860595703, + "learning_rate": 4.82923421338765e-05, + "loss": 1.7814, + "step": 934 + }, + { + "epoch": 0.6557951955111345, + "grad_norm": 0.7961150407791138, + "learning_rate": 4.828707161381423e-05, + "loss": 0.0075, + "step": 935 + }, + { + "epoch": 0.6564965807469753, + "grad_norm": 33.920326232910156, + "learning_rate": 4.8281793261314404e-05, + "loss": 0.7362, + "step": 936 + }, + { + "epoch": 0.6571979659828161, + "grad_norm": 0.9916706681251526, + "learning_rate": 4.827650707815236e-05, + "loss": 0.0096, + "step": 937 + }, + { + "epoch": 0.6578993512186568, + "grad_norm": 0.7823458909988403, + "learning_rate": 4.827121306610608e-05, + "loss": 0.0074, + "step": 938 + }, + { + "epoch": 0.6586007364544977, + "grad_norm": 0.8576016426086426, + "learning_rate": 4.8265911226956154e-05, + "loss": 0.008, + "step": 939 + }, + { + "epoch": 0.6593021216903384, + "grad_norm": 0.5242276787757874, + "learning_rate": 4.826060156248582e-05, + "loss": 0.0047, + "step": 940 + }, + { + "epoch": 0.6600035069261792, + "grad_norm": 0.31952106952667236, + "learning_rate": 4.825528407448096e-05, + "loss": 0.0028, + "step": 941 + }, + { + "epoch": 0.66070489216202, + "grad_norm": 0.17970554530620575, + "learning_rate": 4.824995876473004e-05, + "loss": 0.0015, + "step": 942 + }, + { + "epoch": 0.6614062773978607, + "grad_norm": 0.10021814703941345, + "learning_rate": 4.824462563502422e-05, + "loss": 0.0008, + "step": 943 + }, + { + "epoch": 0.6621076626337016, + "grad_norm": 146.59963989257812, + "learning_rate": 4.823928468715724e-05, + "loss": 5.6442, + "step": 944 + }, + { + "epoch": 0.6628090478695423, + "grad_norm": 0.060028087347745895, + "learning_rate": 4.82339359229255e-05, + "loss": 0.0005, + "step": 945 + }, + { + "epoch": 0.6635104331053832, + "grad_norm": 0.05350363254547119, + "learning_rate": 4.822857934412802e-05, + "loss": 0.0004, + "step": 946 + }, + { + "epoch": 0.6642118183412239, + "grad_norm": 197.06178283691406, + "learning_rate": 4.822321495256644e-05, + "loss": 8.2989, + "step": 947 + }, + { + "epoch": 0.6649132035770647, + "grad_norm": 0.047274697571992874, + "learning_rate": 4.821784275004503e-05, + "loss": 0.0004, + "step": 948 + }, + { + "epoch": 0.6656145888129055, + "grad_norm": 0.037716954946517944, + "learning_rate": 4.82124627383707e-05, + "loss": 0.0003, + "step": 949 + }, + { + "epoch": 0.6663159740487463, + "grad_norm": 48.574623107910156, + "learning_rate": 4.8207074919352976e-05, + "loss": 1.7823, + "step": 950 + }, + { + "epoch": 0.667017359284587, + "grad_norm": 0.06911084055900574, + "learning_rate": 4.8201679294804e-05, + "loss": 0.0006, + "step": 951 + }, + { + "epoch": 0.6677187445204279, + "grad_norm": 0.07155507057905197, + "learning_rate": 4.8196275866538574e-05, + "loss": 0.0006, + "step": 952 + }, + { + "epoch": 0.6684201297562686, + "grad_norm": 0.10516396909952164, + "learning_rate": 4.8190864636374086e-05, + "loss": 0.0009, + "step": 953 + }, + { + "epoch": 0.6691215149921094, + "grad_norm": 46.50364685058594, + "learning_rate": 4.818544560613056e-05, + "loss": 1.4989, + "step": 954 + }, + { + "epoch": 0.6698229002279502, + "grad_norm": 0.2198321670293808, + "learning_rate": 4.8180018777630664e-05, + "loss": 0.0019, + "step": 955 + }, + { + "epoch": 0.670524285463791, + "grad_norm": 87.00233459472656, + "learning_rate": 4.8174584152699665e-05, + "loss": 2.3952, + "step": 956 + }, + { + "epoch": 0.6712256706996318, + "grad_norm": 39.932960510253906, + "learning_rate": 4.816914173316545e-05, + "loss": 1.0167, + "step": 957 + }, + { + "epoch": 0.6719270559354725, + "grad_norm": 0.8517984747886658, + "learning_rate": 4.816369152085855e-05, + "loss": 0.0082, + "step": 958 + }, + { + "epoch": 0.6726284411713134, + "grad_norm": 1.0399830341339111, + "learning_rate": 4.8158233517612096e-05, + "loss": 0.0102, + "step": 959 + }, + { + "epoch": 0.6733298264071541, + "grad_norm": 1.2987735271453857, + "learning_rate": 4.815276772526186e-05, + "loss": 0.013, + "step": 960 + }, + { + "epoch": 0.674031211642995, + "grad_norm": 1.1722874641418457, + "learning_rate": 4.8147294145646207e-05, + "loss": 0.0119, + "step": 961 + }, + { + "epoch": 0.6747325968788357, + "grad_norm": 30.018028259277344, + "learning_rate": 4.8141812780606137e-05, + "loss": 0.606, + "step": 962 + }, + { + "epoch": 0.6754339821146765, + "grad_norm": 1.0533334016799927, + "learning_rate": 4.813632363198527e-05, + "loss": 0.0105, + "step": 963 + }, + { + "epoch": 0.6761353673505173, + "grad_norm": 52.103050231933594, + "learning_rate": 4.813082670162986e-05, + "loss": 1.5312, + "step": 964 + }, + { + "epoch": 0.676836752586358, + "grad_norm": 36.39682388305664, + "learning_rate": 4.812532199138873e-05, + "loss": 0.8569, + "step": 965 + }, + { + "epoch": 0.6775381378221988, + "grad_norm": 84.3869400024414, + "learning_rate": 4.8119809503113363e-05, + "loss": 2.1346, + "step": 966 + }, + { + "epoch": 0.6782395230580396, + "grad_norm": 0.8813936710357666, + "learning_rate": 4.811428923865785e-05, + "loss": 0.0085, + "step": 967 + }, + { + "epoch": 0.6789409082938804, + "grad_norm": 1.1427128314971924, + "learning_rate": 4.810876119987888e-05, + "loss": 0.0113, + "step": 968 + }, + { + "epoch": 0.6796422935297212, + "grad_norm": 30.400836944580078, + "learning_rate": 4.810322538863578e-05, + "loss": 0.6341, + "step": 969 + }, + { + "epoch": 0.680343678765562, + "grad_norm": 77.0494155883789, + "learning_rate": 4.8097681806790466e-05, + "loss": 1.7974, + "step": 970 + }, + { + "epoch": 0.6810450640014027, + "grad_norm": 29.669004440307617, + "learning_rate": 4.8092130456207495e-05, + "loss": 0.5951, + "step": 971 + }, + { + "epoch": 0.6817464492372436, + "grad_norm": 2.0304183959960938, + "learning_rate": 4.808657133875402e-05, + "loss": 0.0214, + "step": 972 + }, + { + "epoch": 0.6824478344730843, + "grad_norm": 22.323291778564453, + "learning_rate": 4.808100445629981e-05, + "loss": 0.4383, + "step": 973 + }, + { + "epoch": 0.6831492197089252, + "grad_norm": 11.758098602294922, + "learning_rate": 4.807542981071724e-05, + "loss": 0.245, + "step": 974 + }, + { + "epoch": 0.6838506049447659, + "grad_norm": 9.657792091369629, + "learning_rate": 4.80698474038813e-05, + "loss": 0.2276, + "step": 975 + }, + { + "epoch": 0.6845519901806068, + "grad_norm": 0.9010874629020691, + "learning_rate": 4.806425723766961e-05, + "loss": 0.1934, + "step": 976 + }, + { + "epoch": 0.6852533754164475, + "grad_norm": 13.536153793334961, + "learning_rate": 4.805865931396236e-05, + "loss": 0.2465, + "step": 977 + }, + { + "epoch": 0.6859547606522882, + "grad_norm": 12.555607795715332, + "learning_rate": 4.8053053634642384e-05, + "loss": 0.2135, + "step": 978 + }, + { + "epoch": 0.6866561458881291, + "grad_norm": 15.943707466125488, + "learning_rate": 4.804744020159511e-05, + "loss": 0.3215, + "step": 979 + }, + { + "epoch": 0.6873575311239698, + "grad_norm": 9.25063419342041, + "learning_rate": 4.804181901670856e-05, + "loss": 0.2798, + "step": 980 + }, + { + "epoch": 0.6880589163598106, + "grad_norm": 8.65345573425293, + "learning_rate": 4.80361900818734e-05, + "loss": 0.1354, + "step": 981 + }, + { + "epoch": 0.6887603015956514, + "grad_norm": 10.924114227294922, + "learning_rate": 4.803055339898288e-05, + "loss": 0.1829, + "step": 982 + }, + { + "epoch": 0.6894616868314922, + "grad_norm": 7.537076950073242, + "learning_rate": 4.8024908969932844e-05, + "loss": 0.1064, + "step": 983 + }, + { + "epoch": 0.690163072067333, + "grad_norm": 43.258338928222656, + "learning_rate": 4.8019256796621764e-05, + "loss": 0.957, + "step": 984 + }, + { + "epoch": 0.6908644573031738, + "grad_norm": 26.308055877685547, + "learning_rate": 4.801359688095071e-05, + "loss": 0.541, + "step": 985 + }, + { + "epoch": 0.6915658425390145, + "grad_norm": 1.8512808084487915, + "learning_rate": 4.800792922482333e-05, + "loss": 0.0192, + "step": 986 + }, + { + "epoch": 0.6922672277748554, + "grad_norm": 1.0505925416946411, + "learning_rate": 4.800225383014594e-05, + "loss": 0.0104, + "step": 987 + }, + { + "epoch": 0.6929686130106961, + "grad_norm": 40.21651840209961, + "learning_rate": 4.799657069882738e-05, + "loss": 1.0056, + "step": 988 + }, + { + "epoch": 0.6936699982465369, + "grad_norm": 0.17625601589679718, + "learning_rate": 4.7990879832779156e-05, + "loss": 0.0015, + "step": 989 + }, + { + "epoch": 0.6943713834823777, + "grad_norm": 0.15104326605796814, + "learning_rate": 4.7985181233915335e-05, + "loss": 0.0013, + "step": 990 + }, + { + "epoch": 0.6950727687182184, + "grad_norm": 0.09187794476747513, + "learning_rate": 4.79794749041526e-05, + "loss": 0.0008, + "step": 991 + }, + { + "epoch": 0.6957741539540593, + "grad_norm": 48.85366439819336, + "learning_rate": 4.797376084541024e-05, + "loss": 1.9766, + "step": 992 + }, + { + "epoch": 0.6964755391899, + "grad_norm": 0.04326195642352104, + "learning_rate": 4.796803905961013e-05, + "loss": 0.0004, + "step": 993 + }, + { + "epoch": 0.6971769244257409, + "grad_norm": 0.04548042640089989, + "learning_rate": 4.796230954867676e-05, + "loss": 0.0004, + "step": 994 + }, + { + "epoch": 0.6978783096615816, + "grad_norm": 48.174827575683594, + "learning_rate": 4.79565723145372e-05, + "loss": 1.9204, + "step": 995 + }, + { + "epoch": 0.6985796948974224, + "grad_norm": 0.04808010160923004, + "learning_rate": 4.7950827359121135e-05, + "loss": 0.0004, + "step": 996 + }, + { + "epoch": 0.6992810801332632, + "grad_norm": 48.49928665161133, + "learning_rate": 4.794507468436084e-05, + "loss": 1.8132, + "step": 997 + }, + { + "epoch": 0.699982465369104, + "grad_norm": 47.405731201171875, + "learning_rate": 4.793931429219117e-05, + "loss": 1.6124, + "step": 998 + }, + { + "epoch": 0.7006838506049448, + "grad_norm": 0.2506006062030792, + "learning_rate": 4.793354618454961e-05, + "loss": 0.0022, + "step": 999 + }, + { + "epoch": 0.7013852358407856, + "grad_norm": 90.9857177734375, + "learning_rate": 4.792777036337621e-05, + "loss": 2.7266, + "step": 1000 + }, + { + "epoch": 0.7020866210766263, + "grad_norm": 0.9927763938903809, + "learning_rate": 4.792198683061363e-05, + "loss": 0.0099, + "step": 1001 + }, + { + "epoch": 0.7027880063124671, + "grad_norm": 1.5031591653823853, + "learning_rate": 4.791619558820712e-05, + "loss": 0.0152, + "step": 1002 + }, + { + "epoch": 0.7034893915483079, + "grad_norm": 29.284543991088867, + "learning_rate": 4.791039663810452e-05, + "loss": 0.5838, + "step": 1003 + }, + { + "epoch": 0.7041907767841487, + "grad_norm": 1.6648647785186768, + "learning_rate": 4.790458998225627e-05, + "loss": 0.0173, + "step": 1004 + }, + { + "epoch": 0.7048921620199895, + "grad_norm": 2.1368696689605713, + "learning_rate": 4.7898775622615384e-05, + "loss": 0.023, + "step": 1005 + }, + { + "epoch": 0.7055935472558302, + "grad_norm": 1.5531522035598755, + "learning_rate": 4.789295356113749e-05, + "loss": 0.0161, + "step": 1006 + }, + { + "epoch": 0.7062949324916711, + "grad_norm": 1.0346095561981201, + "learning_rate": 4.7887123799780795e-05, + "loss": 0.0096, + "step": 1007 + }, + { + "epoch": 0.7069963177275118, + "grad_norm": 0.5431362986564636, + "learning_rate": 4.788128634050609e-05, + "loss": 0.005, + "step": 1008 + }, + { + "epoch": 0.7076977029633527, + "grad_norm": 0.19359983503818512, + "learning_rate": 4.787544118527677e-05, + "loss": 0.0017, + "step": 1009 + }, + { + "epoch": 0.7083990881991934, + "grad_norm": 49.78437042236328, + "learning_rate": 4.786958833605882e-05, + "loss": 1.7876, + "step": 1010 + }, + { + "epoch": 0.7091004734350342, + "grad_norm": 57.95826721191406, + "learning_rate": 4.786372779482078e-05, + "loss": 2.9286, + "step": 1011 + }, + { + "epoch": 0.709801858670875, + "grad_norm": 48.60955047607422, + "learning_rate": 4.7857859563533814e-05, + "loss": 1.9456, + "step": 1012 + }, + { + "epoch": 0.7105032439067157, + "grad_norm": 48.62706756591797, + "learning_rate": 4.785198364417166e-05, + "loss": 1.8589, + "step": 1013 + }, + { + "epoch": 0.7112046291425566, + "grad_norm": 94.03495788574219, + "learning_rate": 4.784610003871063e-05, + "loss": 3.2697, + "step": 1014 + }, + { + "epoch": 0.7119060143783973, + "grad_norm": 44.21165084838867, + "learning_rate": 4.784020874912965e-05, + "loss": 1.3436, + "step": 1015 + }, + { + "epoch": 0.7126073996142381, + "grad_norm": 0.36148276925086975, + "learning_rate": 4.7834309777410195e-05, + "loss": 0.0033, + "step": 1016 + }, + { + "epoch": 0.7133087848500789, + "grad_norm": 74.5863265991211, + "learning_rate": 4.7828403125536354e-05, + "loss": 1.8025, + "step": 1017 + }, + { + "epoch": 0.7140101700859197, + "grad_norm": 43.957252502441406, + "learning_rate": 4.782248879549478e-05, + "loss": 1.1635, + "step": 1018 + }, + { + "epoch": 0.7147115553217604, + "grad_norm": 69.43833923339844, + "learning_rate": 4.781656678927471e-05, + "loss": 1.2, + "step": 1019 + }, + { + "epoch": 0.7154129405576013, + "grad_norm": 4.946931838989258, + "learning_rate": 4.7810637108867986e-05, + "loss": 0.0617, + "step": 1020 + }, + { + "epoch": 0.716114325793442, + "grad_norm": 2.806675434112549, + "learning_rate": 4.7804699756268987e-05, + "loss": 0.1936, + "step": 1021 + }, + { + "epoch": 0.7168157110292829, + "grad_norm": 10.467972755432129, + "learning_rate": 4.7798754733474724e-05, + "loss": 0.17, + "step": 1022 + }, + { + "epoch": 0.7175170962651236, + "grad_norm": 13.816707611083984, + "learning_rate": 4.7792802042484746e-05, + "loss": 0.2481, + "step": 1023 + }, + { + "epoch": 0.7182184815009645, + "grad_norm": 14.67568588256836, + "learning_rate": 4.77868416853012e-05, + "loss": 0.2728, + "step": 1024 + }, + { + "epoch": 0.7189198667368052, + "grad_norm": 3.8597590923309326, + "learning_rate": 4.778087366392881e-05, + "loss": 0.3473, + "step": 1025 + }, + { + "epoch": 0.7196212519726459, + "grad_norm": 6.207676410675049, + "learning_rate": 4.777489798037489e-05, + "loss": 0.276, + "step": 1026 + }, + { + "epoch": 0.7203226372084868, + "grad_norm": 1.4748679399490356, + "learning_rate": 4.77689146366493e-05, + "loss": 0.2487, + "step": 1027 + }, + { + "epoch": 0.7210240224443275, + "grad_norm": 3.2220802307128906, + "learning_rate": 4.77629236347645e-05, + "loss": 0.1941, + "step": 1028 + }, + { + "epoch": 0.7217254076801684, + "grad_norm": 8.897248268127441, + "learning_rate": 4.775692497673552e-05, + "loss": 0.135, + "step": 1029 + }, + { + "epoch": 0.7224267929160091, + "grad_norm": 6.12791633605957, + "learning_rate": 4.7750918664579956e-05, + "loss": 0.0784, + "step": 1030 + }, + { + "epoch": 0.7231281781518499, + "grad_norm": 20.013917922973633, + "learning_rate": 4.7744904700318006e-05, + "loss": 0.3768, + "step": 1031 + }, + { + "epoch": 0.7238295633876907, + "grad_norm": 2.475820541381836, + "learning_rate": 4.7738883085972406e-05, + "loss": 0.0274, + "step": 1032 + }, + { + "epoch": 0.7245309486235315, + "grad_norm": 1.4534077644348145, + "learning_rate": 4.773285382356849e-05, + "loss": 0.015, + "step": 1033 + }, + { + "epoch": 0.7252323338593722, + "grad_norm": 36.22199630737305, + "learning_rate": 4.772681691513417e-05, + "loss": 0.9046, + "step": 1034 + }, + { + "epoch": 0.7259337190952131, + "grad_norm": 0.2810744643211365, + "learning_rate": 4.772077236269989e-05, + "loss": 0.0025, + "step": 1035 + }, + { + "epoch": 0.7266351043310538, + "grad_norm": 45.44533920288086, + "learning_rate": 4.771472016829871e-05, + "loss": 1.3375, + "step": 1036 + }, + { + "epoch": 0.7273364895668946, + "grad_norm": 0.12535035610198975, + "learning_rate": 4.770866033396623e-05, + "loss": 0.0011, + "step": 1037 + }, + { + "epoch": 0.7280378748027354, + "grad_norm": 45.64175033569336, + "learning_rate": 4.770259286174065e-05, + "loss": 1.6729, + "step": 1038 + }, + { + "epoch": 0.7287392600385761, + "grad_norm": 46.727394104003906, + "learning_rate": 4.7696517753662696e-05, + "loss": 1.7577, + "step": 1039 + }, + { + "epoch": 0.729440645274417, + "grad_norm": 0.17975449562072754, + "learning_rate": 4.76904350117757e-05, + "loss": 0.0016, + "step": 1040 + }, + { + "epoch": 0.7301420305102577, + "grad_norm": 85.70226287841797, + "learning_rate": 4.7684344638125546e-05, + "loss": 2.662, + "step": 1041 + }, + { + "epoch": 0.7308434157460986, + "grad_norm": 39.62570571899414, + "learning_rate": 4.767824663476069e-05, + "loss": 1.1086, + "step": 1042 + }, + { + "epoch": 0.7315448009819393, + "grad_norm": 33.90923309326172, + "learning_rate": 4.767214100373214e-05, + "loss": 0.8021, + "step": 1043 + }, + { + "epoch": 0.7322461862177801, + "grad_norm": 23.501230239868164, + "learning_rate": 4.76660277470935e-05, + "loss": 0.4899, + "step": 1044 + }, + { + "epoch": 0.7329475714536209, + "grad_norm": 10.125417709350586, + "learning_rate": 4.76599068669009e-05, + "loss": 0.2671, + "step": 1045 + }, + { + "epoch": 0.7336489566894617, + "grad_norm": 3.0318775177001953, + "learning_rate": 4.7653778365213076e-05, + "loss": 0.2053, + "step": 1046 + }, + { + "epoch": 0.7343503419253025, + "grad_norm": 7.598801612854004, + "learning_rate": 4.7647642244091276e-05, + "loss": 0.2937, + "step": 1047 + }, + { + "epoch": 0.7350517271611433, + "grad_norm": 11.337316513061523, + "learning_rate": 4.7641498505599356e-05, + "loss": 0.283, + "step": 1048 + }, + { + "epoch": 0.735753112396984, + "grad_norm": 2.909923553466797, + "learning_rate": 4.763534715180373e-05, + "loss": 0.2887, + "step": 1049 + }, + { + "epoch": 0.7364544976328248, + "grad_norm": 12.51762866973877, + "learning_rate": 4.762918818477334e-05, + "loss": 0.2306, + "step": 1050 + }, + { + "epoch": 0.7371558828686656, + "grad_norm": 10.550107955932617, + "learning_rate": 4.762302160657971e-05, + "loss": 0.1879, + "step": 1051 + }, + { + "epoch": 0.7378572681045064, + "grad_norm": 10.876016616821289, + "learning_rate": 4.7616847419296945e-05, + "loss": 0.1877, + "step": 1052 + }, + { + "epoch": 0.7385586533403472, + "grad_norm": 3.0670106410980225, + "learning_rate": 4.761066562500167e-05, + "loss": 0.2327, + "step": 1053 + }, + { + "epoch": 0.7392600385761879, + "grad_norm": 5.0195770263671875, + "learning_rate": 4.7604476225773084e-05, + "loss": 0.0651, + "step": 1054 + }, + { + "epoch": 0.7399614238120288, + "grad_norm": 2.679926872253418, + "learning_rate": 4.759827922369296e-05, + "loss": 0.0317, + "step": 1055 + }, + { + "epoch": 0.7406628090478695, + "grad_norm": 1.752316951751709, + "learning_rate": 4.7592074620845616e-05, + "loss": 0.0184, + "step": 1056 + }, + { + "epoch": 0.7413641942837104, + "grad_norm": 0.6544418334960938, + "learning_rate": 4.758586241931791e-05, + "loss": 0.0062, + "step": 1057 + }, + { + "epoch": 0.7420655795195511, + "grad_norm": 0.19106708467006683, + "learning_rate": 4.757964262119928e-05, + "loss": 0.0017, + "step": 1058 + }, + { + "epoch": 0.742766964755392, + "grad_norm": 0.0741598978638649, + "learning_rate": 4.757341522858171e-05, + "loss": 0.0006, + "step": 1059 + }, + { + "epoch": 0.7434683499912327, + "grad_norm": 0.02510019950568676, + "learning_rate": 4.7567180243559736e-05, + "loss": 0.0002, + "step": 1060 + }, + { + "epoch": 0.7441697352270735, + "grad_norm": 0.00887396652251482, + "learning_rate": 4.7560937668230454e-05, + "loss": 0.0001, + "step": 1061 + }, + { + "epoch": 0.7448711204629143, + "grad_norm": 0.0027078278362751007, + "learning_rate": 4.75546875046935e-05, + "loss": 0.0, + "step": 1062 + }, + { + "epoch": 0.745572505698755, + "grad_norm": 0.001226524356752634, + "learning_rate": 4.7548429755051085e-05, + "loss": 0.0, + "step": 1063 + }, + { + "epoch": 0.7462738909345958, + "grad_norm": 0.0005389907164499164, + "learning_rate": 4.754216442140794e-05, + "loss": 0.0, + "step": 1064 + }, + { + "epoch": 0.7469752761704366, + "grad_norm": 55.77132797241211, + "learning_rate": 4.7535891505871375e-05, + "loss": 3.9293, + "step": 1065 + }, + { + "epoch": 0.7476766614062774, + "grad_norm": 103.8635025024414, + "learning_rate": 4.752961101055124e-05, + "loss": 7.9354, + "step": 1066 + }, + { + "epoch": 0.7483780466421182, + "grad_norm": 103.83565521240234, + "learning_rate": 4.7523322937559946e-05, + "loss": 7.4002, + "step": 1067 + }, + { + "epoch": 0.749079431877959, + "grad_norm": 0.0008341918000951409, + "learning_rate": 4.751702728901242e-05, + "loss": 0.0, + "step": 1068 + }, + { + "epoch": 0.7497808171137997, + "grad_norm": 48.793819427490234, + "learning_rate": 4.7510724067026156e-05, + "loss": 3.1709, + "step": 1069 + }, + { + "epoch": 0.7504822023496406, + "grad_norm": 97.3034439086914, + "learning_rate": 4.750441327372122e-05, + "loss": 5.7, + "step": 1070 + }, + { + "epoch": 0.7511835875854813, + "grad_norm": 0.0208453219383955, + "learning_rate": 4.7498094911220185e-05, + "loss": 0.0002, + "step": 1071 + }, + { + "epoch": 0.7518849728213222, + "grad_norm": 0.05400345101952553, + "learning_rate": 4.749176898164818e-05, + "loss": 0.0004, + "step": 1072 + }, + { + "epoch": 0.7525863580571629, + "grad_norm": 0.0927615538239479, + "learning_rate": 4.7485435487132904e-05, + "loss": 0.0008, + "step": 1073 + }, + { + "epoch": 0.7532877432930036, + "grad_norm": 0.19018051028251648, + "learning_rate": 4.747909442980456e-05, + "loss": 0.0016, + "step": 1074 + }, + { + "epoch": 0.7539891285288445, + "grad_norm": 0.3820917010307312, + "learning_rate": 4.7472745811795936e-05, + "loss": 0.0035, + "step": 1075 + }, + { + "epoch": 0.7546905137646852, + "grad_norm": 0.6652111411094666, + "learning_rate": 4.746638963524235e-05, + "loss": 0.0063, + "step": 1076 + }, + { + "epoch": 0.7553918990005261, + "grad_norm": 91.62269592285156, + "learning_rate": 4.7460025902281625e-05, + "loss": 2.2848, + "step": 1077 + }, + { + "epoch": 0.7560932842363668, + "grad_norm": 0.6401788592338562, + "learning_rate": 4.7453654615054174e-05, + "loss": 0.006, + "step": 1078 + }, + { + "epoch": 0.7567946694722076, + "grad_norm": 0.9835159182548523, + "learning_rate": 4.744727577570295e-05, + "loss": 0.0096, + "step": 1079 + }, + { + "epoch": 0.7574960547080484, + "grad_norm": 0.665366530418396, + "learning_rate": 4.74408893863734e-05, + "loss": 0.0064, + "step": 1080 + }, + { + "epoch": 0.7581974399438892, + "grad_norm": 0.3792603313922882, + "learning_rate": 4.743449544921356e-05, + "loss": 0.0034, + "step": 1081 + }, + { + "epoch": 0.75889882517973, + "grad_norm": 0.18616673350334167, + "learning_rate": 4.742809396637397e-05, + "loss": 0.0016, + "step": 1082 + }, + { + "epoch": 0.7596002104155708, + "grad_norm": 0.12753531336784363, + "learning_rate": 4.7421684940007725e-05, + "loss": 0.0011, + "step": 1083 + }, + { + "epoch": 0.7603015956514115, + "grad_norm": 92.24988555908203, + "learning_rate": 4.741526837227047e-05, + "loss": 3.521, + "step": 1084 + }, + { + "epoch": 0.7610029808872524, + "grad_norm": 48.01845169067383, + "learning_rate": 4.740884426532035e-05, + "loss": 1.9461, + "step": 1085 + }, + { + "epoch": 0.7617043661230931, + "grad_norm": 0.07205016911029816, + "learning_rate": 4.740241262131808e-05, + "loss": 0.0006, + "step": 1086 + }, + { + "epoch": 0.7624057513589338, + "grad_norm": 0.07788128405809402, + "learning_rate": 4.7395973442426895e-05, + "loss": 0.0007, + "step": 1087 + }, + { + "epoch": 0.7631071365947747, + "grad_norm": 94.46582794189453, + "learning_rate": 4.738952673081256e-05, + "loss": 3.4385, + "step": 1088 + }, + { + "epoch": 0.7638085218306154, + "grad_norm": 0.17950232326984406, + "learning_rate": 4.738307248864339e-05, + "loss": 0.0016, + "step": 1089 + }, + { + "epoch": 0.7645099070664563, + "grad_norm": 93.51290130615234, + "learning_rate": 4.737661071809021e-05, + "loss": 3.3519, + "step": 1090 + }, + { + "epoch": 0.765211292302297, + "grad_norm": 92.91374206542969, + "learning_rate": 4.73701414213264e-05, + "loss": 2.7943, + "step": 1091 + }, + { + "epoch": 0.7659126775381379, + "grad_norm": 0.5978343486785889, + "learning_rate": 4.736366460052786e-05, + "loss": 0.0056, + "step": 1092 + }, + { + "epoch": 0.7666140627739786, + "grad_norm": 30.616474151611328, + "learning_rate": 4.7357180257873014e-05, + "loss": 0.6588, + "step": 1093 + }, + { + "epoch": 0.7673154480098194, + "grad_norm": 1.5538511276245117, + "learning_rate": 4.7350688395542825e-05, + "loss": 0.0162, + "step": 1094 + }, + { + "epoch": 0.7680168332456602, + "grad_norm": 75.00251007080078, + "learning_rate": 4.734418901572079e-05, + "loss": 1.4034, + "step": 1095 + }, + { + "epoch": 0.768718218481501, + "grad_norm": 3.7650375366210938, + "learning_rate": 4.733768212059293e-05, + "loss": 0.0454, + "step": 1096 + }, + { + "epoch": 0.7694196037173417, + "grad_norm": 5.952811241149902, + "learning_rate": 4.733116771234778e-05, + "loss": 0.0766, + "step": 1097 + }, + { + "epoch": 0.7701209889531825, + "grad_norm": 15.600150108337402, + "learning_rate": 4.7324645793176425e-05, + "loss": 0.3215, + "step": 1098 + }, + { + "epoch": 0.7708223741890233, + "grad_norm": 9.213759422302246, + "learning_rate": 4.731811636527246e-05, + "loss": 0.2548, + "step": 1099 + }, + { + "epoch": 0.7715237594248641, + "grad_norm": 7.49055814743042, + "learning_rate": 4.731157943083202e-05, + "loss": 0.1, + "step": 1100 + }, + { + "epoch": 0.7722251446607049, + "grad_norm": 7.580368518829346, + "learning_rate": 4.730503499205373e-05, + "loss": 0.1067, + "step": 1101 + }, + { + "epoch": 0.7729265298965456, + "grad_norm": 6.691344261169434, + "learning_rate": 4.7298483051138796e-05, + "loss": 0.0911, + "step": 1102 + }, + { + "epoch": 0.7736279151323865, + "grad_norm": 39.10125732421875, + "learning_rate": 4.72919236102909e-05, + "loss": 0.7558, + "step": 1103 + }, + { + "epoch": 0.7743293003682272, + "grad_norm": 4.285506725311279, + "learning_rate": 4.728535667171626e-05, + "loss": 0.051, + "step": 1104 + }, + { + "epoch": 0.7750306856040681, + "grad_norm": 2.88086199760437, + "learning_rate": 4.727878223762363e-05, + "loss": 0.0324, + "step": 1105 + }, + { + "epoch": 0.7757320708399088, + "grad_norm": 1.6831408739089966, + "learning_rate": 4.7272200310224256e-05, + "loss": 0.0172, + "step": 1106 + }, + { + "epoch": 0.7764334560757497, + "grad_norm": 37.848419189453125, + "learning_rate": 4.726561089173194e-05, + "loss": 0.934, + "step": 1107 + }, + { + "epoch": 0.7771348413115904, + "grad_norm": 38.1249885559082, + "learning_rate": 4.725901398436297e-05, + "loss": 0.9811, + "step": 1108 + }, + { + "epoch": 0.7778362265474312, + "grad_norm": 51.93446731567383, + "learning_rate": 4.725240959033618e-05, + "loss": 1.5657, + "step": 1109 + }, + { + "epoch": 0.778537611783272, + "grad_norm": 0.33258679509162903, + "learning_rate": 4.7245797711872895e-05, + "loss": 0.003, + "step": 1110 + }, + { + "epoch": 0.7792389970191127, + "grad_norm": 61.720924377441406, + "learning_rate": 4.723917835119699e-05, + "loss": 1.6662, + "step": 1111 + }, + { + "epoch": 0.7799403822549535, + "grad_norm": 37.942874908447266, + "learning_rate": 4.723255151053483e-05, + "loss": 1.0046, + "step": 1112 + }, + { + "epoch": 0.7806417674907943, + "grad_norm": 0.8135581016540527, + "learning_rate": 4.7225917192115297e-05, + "loss": 0.0076, + "step": 1113 + }, + { + "epoch": 0.7813431527266351, + "grad_norm": 36.66408920288086, + "learning_rate": 4.721927539816981e-05, + "loss": 0.9099, + "step": 1114 + }, + { + "epoch": 0.7820445379624759, + "grad_norm": 1.1302094459533691, + "learning_rate": 4.721262613093228e-05, + "loss": 0.0115, + "step": 1115 + }, + { + "epoch": 0.7827459231983167, + "grad_norm": 1.4476618766784668, + "learning_rate": 4.720596939263914e-05, + "loss": 0.0146, + "step": 1116 + }, + { + "epoch": 0.7834473084341574, + "grad_norm": 1.0591875314712524, + "learning_rate": 4.719930518552934e-05, + "loss": 0.0106, + "step": 1117 + }, + { + "epoch": 0.7841486936699983, + "grad_norm": 0.7351242899894714, + "learning_rate": 4.7192633511844334e-05, + "loss": 0.0069, + "step": 1118 + }, + { + "epoch": 0.784850078905839, + "grad_norm": 0.43580272793769836, + "learning_rate": 4.71859543738281e-05, + "loss": 0.004, + "step": 1119 + }, + { + "epoch": 0.7855514641416799, + "grad_norm": 0.256906658411026, + "learning_rate": 4.71792677737271e-05, + "loss": 0.0022, + "step": 1120 + }, + { + "epoch": 0.7862528493775206, + "grad_norm": 0.13365890085697174, + "learning_rate": 4.717257371379034e-05, + "loss": 0.0012, + "step": 1121 + }, + { + "epoch": 0.7869542346133613, + "grad_norm": 0.07990420609712601, + "learning_rate": 4.71658721962693e-05, + "loss": 0.0007, + "step": 1122 + }, + { + "epoch": 0.7876556198492022, + "grad_norm": 0.0479491651058197, + "learning_rate": 4.715916322341801e-05, + "loss": 0.0004, + "step": 1123 + }, + { + "epoch": 0.7883570050850429, + "grad_norm": 47.28627014160156, + "learning_rate": 4.7152446797492963e-05, + "loss": 2.0982, + "step": 1124 + }, + { + "epoch": 0.7890583903208838, + "grad_norm": 96.63157653808594, + "learning_rate": 4.71457229207532e-05, + "loss": 4.5257, + "step": 1125 + }, + { + "epoch": 0.7897597755567245, + "grad_norm": 47.81488800048828, + "learning_rate": 4.713899159546024e-05, + "loss": 2.3268, + "step": 1126 + }, + { + "epoch": 0.7904611607925653, + "grad_norm": 47.34000015258789, + "learning_rate": 4.713225282387811e-05, + "loss": 2.1484, + "step": 1127 + }, + { + "epoch": 0.7911625460284061, + "grad_norm": 46.76436996459961, + "learning_rate": 4.7125506608273344e-05, + "loss": 1.8822, + "step": 1128 + }, + { + "epoch": 0.7918639312642469, + "grad_norm": 0.12850354611873627, + "learning_rate": 4.711875295091499e-05, + "loss": 0.0011, + "step": 1129 + }, + { + "epoch": 0.7925653165000877, + "grad_norm": 126.00606536865234, + "learning_rate": 4.71119918540746e-05, + "loss": 3.7093, + "step": 1130 + }, + { + "epoch": 0.7932667017359285, + "grad_norm": 0.4526232182979584, + "learning_rate": 4.7105223320026205e-05, + "loss": 0.0042, + "step": 1131 + }, + { + "epoch": 0.7939680869717692, + "grad_norm": 33.449928283691406, + "learning_rate": 4.709844735104636e-05, + "loss": 0.7578, + "step": 1132 + }, + { + "epoch": 0.7946694722076101, + "grad_norm": 1.6459629535675049, + "learning_rate": 4.7091663949414105e-05, + "loss": 0.0174, + "step": 1133 + }, + { + "epoch": 0.7953708574434508, + "grad_norm": 34.26624298095703, + "learning_rate": 4.7084873117411e-05, + "loss": 0.7867, + "step": 1134 + }, + { + "epoch": 0.7960722426792916, + "grad_norm": 3.238154888153076, + "learning_rate": 4.7078074857321085e-05, + "loss": 0.0382, + "step": 1135 + }, + { + "epoch": 0.7967736279151324, + "grad_norm": 3.573115348815918, + "learning_rate": 4.70712691714309e-05, + "loss": 0.0431, + "step": 1136 + }, + { + "epoch": 0.7974750131509731, + "grad_norm": 4.002458095550537, + "learning_rate": 4.7064456062029506e-05, + "loss": 0.0488, + "step": 1137 + }, + { + "epoch": 0.798176398386814, + "grad_norm": 3.4627814292907715, + "learning_rate": 4.7057635531408424e-05, + "loss": 0.0394, + "step": 1138 + }, + { + "epoch": 0.7988777836226547, + "grad_norm": 18.19850730895996, + "learning_rate": 4.70508075818617e-05, + "loss": 0.3679, + "step": 1139 + }, + { + "epoch": 0.7995791688584956, + "grad_norm": 2.7357804775238037, + "learning_rate": 4.704397221568586e-05, + "loss": 0.0312, + "step": 1140 + }, + { + "epoch": 0.8002805540943363, + "grad_norm": 51.47498321533203, + "learning_rate": 4.703712943517993e-05, + "loss": 1.0259, + "step": 1141 + }, + { + "epoch": 0.8009819393301771, + "grad_norm": 1.7935067415237427, + "learning_rate": 4.7030279242645434e-05, + "loss": 0.0192, + "step": 1142 + }, + { + "epoch": 0.8016833245660179, + "grad_norm": 28.79102325439453, + "learning_rate": 4.702342164038639e-05, + "loss": 0.6393, + "step": 1143 + }, + { + "epoch": 0.8023847098018587, + "grad_norm": 1.230648398399353, + "learning_rate": 4.7016556630709284e-05, + "loss": 0.0126, + "step": 1144 + }, + { + "epoch": 0.8030860950376995, + "grad_norm": 0.9287470579147339, + "learning_rate": 4.700968421592313e-05, + "loss": 0.0094, + "step": 1145 + }, + { + "epoch": 0.8037874802735402, + "grad_norm": 62.944549560546875, + "learning_rate": 4.700280439833941e-05, + "loss": 1.3975, + "step": 1146 + }, + { + "epoch": 0.804488865509381, + "grad_norm": 44.87819290161133, + "learning_rate": 4.69959171802721e-05, + "loss": 1.0857, + "step": 1147 + }, + { + "epoch": 0.8051902507452218, + "grad_norm": 0.5283865928649902, + "learning_rate": 4.6989022564037654e-05, + "loss": 0.0049, + "step": 1148 + }, + { + "epoch": 0.8058916359810626, + "grad_norm": 0.5992218852043152, + "learning_rate": 4.698212055195505e-05, + "loss": 0.0058, + "step": 1149 + }, + { + "epoch": 0.8065930212169033, + "grad_norm": 0.5411283373832703, + "learning_rate": 4.697521114634571e-05, + "loss": 0.0052, + "step": 1150 + }, + { + "epoch": 0.8072944064527442, + "grad_norm": 75.1710433959961, + "learning_rate": 4.696829434953357e-05, + "loss": 1.9684, + "step": 1151 + }, + { + "epoch": 0.8079957916885849, + "grad_norm": 0.3737725615501404, + "learning_rate": 4.6961370163845044e-05, + "loss": 0.0035, + "step": 1152 + }, + { + "epoch": 0.8086971769244258, + "grad_norm": 48.369232177734375, + "learning_rate": 4.6954438591609027e-05, + "loss": 1.2948, + "step": 1153 + }, + { + "epoch": 0.8093985621602665, + "grad_norm": 0.4170537292957306, + "learning_rate": 4.694749963515691e-05, + "loss": 0.0036, + "step": 1154 + }, + { + "epoch": 0.8100999473961074, + "grad_norm": 0.3476961553096771, + "learning_rate": 4.6940553296822565e-05, + "loss": 0.0031, + "step": 1155 + }, + { + "epoch": 0.8108013326319481, + "grad_norm": 40.558204650878906, + "learning_rate": 4.693359957894232e-05, + "loss": 1.199, + "step": 1156 + }, + { + "epoch": 0.8115027178677889, + "grad_norm": 78.08576202392578, + "learning_rate": 4.692663848385503e-05, + "loss": 2.1697, + "step": 1157 + }, + { + "epoch": 0.8122041031036297, + "grad_norm": 0.5332819819450378, + "learning_rate": 4.6919670013902015e-05, + "loss": 0.0051, + "step": 1158 + }, + { + "epoch": 0.8129054883394704, + "grad_norm": 0.6460719108581543, + "learning_rate": 4.691269417142704e-05, + "loss": 0.0063, + "step": 1159 + }, + { + "epoch": 0.8136068735753113, + "grad_norm": 0.6999556422233582, + "learning_rate": 4.69057109587764e-05, + "loss": 0.0068, + "step": 1160 + }, + { + "epoch": 0.814308258811152, + "grad_norm": 0.5912299156188965, + "learning_rate": 4.6898720378298846e-05, + "loss": 0.0057, + "step": 1161 + }, + { + "epoch": 0.8150096440469928, + "grad_norm": 0.4620411992073059, + "learning_rate": 4.6891722432345597e-05, + "loss": 0.0044, + "step": 1162 + }, + { + "epoch": 0.8157110292828336, + "grad_norm": 0.3710283935070038, + "learning_rate": 4.6884717123270374e-05, + "loss": 0.0034, + "step": 1163 + }, + { + "epoch": 0.8164124145186744, + "grad_norm": 0.22289921343326569, + "learning_rate": 4.687770445342936e-05, + "loss": 0.002, + "step": 1164 + }, + { + "epoch": 0.8171137997545151, + "grad_norm": 89.52765655517578, + "learning_rate": 4.687068442518121e-05, + "loss": 2.8479, + "step": 1165 + }, + { + "epoch": 0.817815184990356, + "grad_norm": 0.15177088975906372, + "learning_rate": 4.6863657040887065e-05, + "loss": 0.0013, + "step": 1166 + }, + { + "epoch": 0.8185165702261967, + "grad_norm": 0.13902179896831512, + "learning_rate": 4.685662230291054e-05, + "loss": 0.0012, + "step": 1167 + }, + { + "epoch": 0.8192179554620376, + "grad_norm": 0.10139818489551544, + "learning_rate": 4.68495802136177e-05, + "loss": 0.0009, + "step": 1168 + }, + { + "epoch": 0.8199193406978783, + "grad_norm": 0.08662685006856918, + "learning_rate": 4.684253077537712e-05, + "loss": 0.0007, + "step": 1169 + }, + { + "epoch": 0.820620725933719, + "grad_norm": 0.07555234432220459, + "learning_rate": 4.683547399055982e-05, + "loss": 0.0006, + "step": 1170 + }, + { + "epoch": 0.8213221111695599, + "grad_norm": 0.06445518881082535, + "learning_rate": 4.682840986153929e-05, + "loss": 0.0006, + "step": 1171 + }, + { + "epoch": 0.8220234964054006, + "grad_norm": 46.72000503540039, + "learning_rate": 4.68213383906915e-05, + "loss": 2.1281, + "step": 1172 + }, + { + "epoch": 0.8227248816412415, + "grad_norm": 46.803985595703125, + "learning_rate": 4.6814259580394915e-05, + "loss": 1.9621, + "step": 1173 + }, + { + "epoch": 0.8234262668770822, + "grad_norm": 0.07187401503324509, + "learning_rate": 4.68071734330304e-05, + "loss": 0.0006, + "step": 1174 + }, + { + "epoch": 0.824127652112923, + "grad_norm": 0.09642617404460907, + "learning_rate": 4.680007995098136e-05, + "loss": 0.0008, + "step": 1175 + }, + { + "epoch": 0.8248290373487638, + "grad_norm": 0.11466558277606964, + "learning_rate": 4.679297913663363e-05, + "loss": 0.001, + "step": 1176 + }, + { + "epoch": 0.8255304225846046, + "grad_norm": 0.11334109306335449, + "learning_rate": 4.67858709923755e-05, + "loss": 0.001, + "step": 1177 + }, + { + "epoch": 0.8262318078204454, + "grad_norm": 0.1963043361902237, + "learning_rate": 4.677875552059777e-05, + "loss": 0.0017, + "step": 1178 + }, + { + "epoch": 0.8269331930562862, + "grad_norm": 0.17632529139518738, + "learning_rate": 4.677163272369365e-05, + "loss": 0.0016, + "step": 1179 + }, + { + "epoch": 0.8276345782921269, + "grad_norm": 124.80657958984375, + "learning_rate": 4.6764502604058864e-05, + "loss": 3.9859, + "step": 1180 + }, + { + "epoch": 0.8283359635279678, + "grad_norm": 0.19632282853126526, + "learning_rate": 4.675736516409156e-05, + "loss": 0.0018, + "step": 1181 + }, + { + "epoch": 0.8290373487638085, + "grad_norm": 43.057350158691406, + "learning_rate": 4.675022040619238e-05, + "loss": 1.2696, + "step": 1182 + }, + { + "epoch": 0.8297387339996493, + "grad_norm": 39.46339416503906, + "learning_rate": 4.674306833276439e-05, + "loss": 1.093, + "step": 1183 + }, + { + "epoch": 0.8304401192354901, + "grad_norm": 0.653581976890564, + "learning_rate": 4.6735908946213156e-05, + "loss": 0.0065, + "step": 1184 + }, + { + "epoch": 0.8311415044713308, + "grad_norm": 0.8602274656295776, + "learning_rate": 4.6728742248946677e-05, + "loss": 0.0088, + "step": 1185 + }, + { + "epoch": 0.8318428897071717, + "grad_norm": 0.9501532912254333, + "learning_rate": 4.6721568243375425e-05, + "loss": 0.0095, + "step": 1186 + }, + { + "epoch": 0.8325442749430124, + "grad_norm": 0.8702390193939209, + "learning_rate": 4.671438693191232e-05, + "loss": 0.0088, + "step": 1187 + }, + { + "epoch": 0.8332456601788533, + "grad_norm": 0.6245111227035522, + "learning_rate": 4.6707198316972746e-05, + "loss": 0.0061, + "step": 1188 + }, + { + "epoch": 0.833947045414694, + "grad_norm": 0.4615590572357178, + "learning_rate": 4.6700002400974544e-05, + "loss": 0.0045, + "step": 1189 + }, + { + "epoch": 0.8346484306505348, + "grad_norm": 0.3482006788253784, + "learning_rate": 4.6692799186338e-05, + "loss": 0.0033, + "step": 1190 + }, + { + "epoch": 0.8353498158863756, + "grad_norm": 43.396461486816406, + "learning_rate": 4.668558867548587e-05, + "loss": 1.3748, + "step": 1191 + }, + { + "epoch": 0.8360512011222164, + "grad_norm": 49.66410827636719, + "learning_rate": 4.6678370870843357e-05, + "loss": 1.7113, + "step": 1192 + }, + { + "epoch": 0.8367525863580572, + "grad_norm": 0.1562856137752533, + "learning_rate": 4.667114577483812e-05, + "loss": 0.0014, + "step": 1193 + }, + { + "epoch": 0.8374539715938979, + "grad_norm": 47.60147476196289, + "learning_rate": 4.666391338990026e-05, + "loss": 1.8113, + "step": 1194 + }, + { + "epoch": 0.8381553568297387, + "grad_norm": 0.19547221064567566, + "learning_rate": 4.665667371846234e-05, + "loss": 0.0018, + "step": 1195 + }, + { + "epoch": 0.8388567420655795, + "grad_norm": 0.23496605455875397, + "learning_rate": 4.664942676295937e-05, + "loss": 0.0022, + "step": 1196 + }, + { + "epoch": 0.8395581273014203, + "grad_norm": 0.19859689474105835, + "learning_rate": 4.6642172525828806e-05, + "loss": 0.0018, + "step": 1197 + }, + { + "epoch": 0.840259512537261, + "grad_norm": 0.21561066806316376, + "learning_rate": 4.6634911009510574e-05, + "loss": 0.002, + "step": 1198 + }, + { + "epoch": 0.8409608977731019, + "grad_norm": 0.16780033707618713, + "learning_rate": 4.662764221644702e-05, + "loss": 0.0016, + "step": 1199 + }, + { + "epoch": 0.8416622830089426, + "grad_norm": 0.170645073056221, + "learning_rate": 4.6620366149082954e-05, + "loss": 0.0015, + "step": 1200 + }, + { + "epoch": 0.8423636682447835, + "grad_norm": 0.16811063885688782, + "learning_rate": 4.661308280986563e-05, + "loss": 0.0015, + "step": 1201 + }, + { + "epoch": 0.8430650534806242, + "grad_norm": 42.41083908081055, + "learning_rate": 4.6605792201244733e-05, + "loss": 1.4838, + "step": 1202 + }, + { + "epoch": 0.8437664387164651, + "grad_norm": 43.9367790222168, + "learning_rate": 4.6598494325672426e-05, + "loss": 1.4149, + "step": 1203 + }, + { + "epoch": 0.8444678239523058, + "grad_norm": 0.1912125051021576, + "learning_rate": 4.659118918560329e-05, + "loss": 0.0018, + "step": 1204 + }, + { + "epoch": 0.8451692091881466, + "grad_norm": 0.27999013662338257, + "learning_rate": 4.6583876783494343e-05, + "loss": 0.0025, + "step": 1205 + }, + { + "epoch": 0.8458705944239874, + "grad_norm": 36.06639099121094, + "learning_rate": 4.657655712180507e-05, + "loss": 0.9947, + "step": 1206 + }, + { + "epoch": 0.8465719796598281, + "grad_norm": 0.4413144588470459, + "learning_rate": 4.6569230202997385e-05, + "loss": 0.0042, + "step": 1207 + }, + { + "epoch": 0.847273364895669, + "grad_norm": 0.530229926109314, + "learning_rate": 4.656189602953565e-05, + "loss": 0.0052, + "step": 1208 + }, + { + "epoch": 0.8479747501315097, + "grad_norm": 0.6561914086341858, + "learning_rate": 4.655455460388665e-05, + "loss": 0.0066, + "step": 1209 + }, + { + "epoch": 0.8486761353673505, + "grad_norm": 0.6459203362464905, + "learning_rate": 4.654720592851962e-05, + "loss": 0.0063, + "step": 1210 + }, + { + "epoch": 0.8493775206031913, + "grad_norm": 37.68455505371094, + "learning_rate": 4.653985000590624e-05, + "loss": 1.0854, + "step": 1211 + }, + { + "epoch": 0.8500789058390321, + "grad_norm": 0.4941748380661011, + "learning_rate": 4.6532486838520616e-05, + "loss": 0.0049, + "step": 1212 + }, + { + "epoch": 0.8507802910748729, + "grad_norm": 38.626068115234375, + "learning_rate": 4.6525116428839307e-05, + "loss": 1.0544, + "step": 1213 + }, + { + "epoch": 0.8514816763107137, + "grad_norm": 0.5088845491409302, + "learning_rate": 4.651773877934129e-05, + "loss": 0.005, + "step": 1214 + }, + { + "epoch": 0.8521830615465544, + "grad_norm": 0.5165444016456604, + "learning_rate": 4.651035389250798e-05, + "loss": 0.005, + "step": 1215 + }, + { + "epoch": 0.8528844467823953, + "grad_norm": 37.2906379699707, + "learning_rate": 4.650296177082323e-05, + "loss": 1.0227, + "step": 1216 + }, + { + "epoch": 0.853585832018236, + "grad_norm": 0.5255476832389832, + "learning_rate": 4.649556241677333e-05, + "loss": 0.0052, + "step": 1217 + }, + { + "epoch": 0.8542872172540767, + "grad_norm": 37.65691375732422, + "learning_rate": 4.6488155832847006e-05, + "loss": 1.1044, + "step": 1218 + }, + { + "epoch": 0.8549886024899176, + "grad_norm": 35.80206298828125, + "learning_rate": 4.648074202153539e-05, + "loss": 0.8649, + "step": 1219 + }, + { + "epoch": 0.8556899877257583, + "grad_norm": 0.8690975308418274, + "learning_rate": 4.647332098533208e-05, + "loss": 0.0089, + "step": 1220 + }, + { + "epoch": 0.8563913729615992, + "grad_norm": 1.0327463150024414, + "learning_rate": 4.646589272673308e-05, + "loss": 0.0109, + "step": 1221 + }, + { + "epoch": 0.8570927581974399, + "grad_norm": 32.14468765258789, + "learning_rate": 4.645845724823684e-05, + "loss": 0.7778, + "step": 1222 + }, + { + "epoch": 0.8577941434332808, + "grad_norm": 1.0274231433868408, + "learning_rate": 4.645101455234422e-05, + "loss": 0.0106, + "step": 1223 + }, + { + "epoch": 0.8584955286691215, + "grad_norm": 29.724233627319336, + "learning_rate": 4.644356464155852e-05, + "loss": 0.6951, + "step": 1224 + }, + { + "epoch": 0.8591969139049623, + "grad_norm": 1.3864684104919434, + "learning_rate": 4.6436107518385465e-05, + "loss": 0.0151, + "step": 1225 + }, + { + "epoch": 0.8598982991408031, + "grad_norm": 1.3250905275344849, + "learning_rate": 4.6428643185333196e-05, + "loss": 0.0144, + "step": 1226 + }, + { + "epoch": 0.8605996843766439, + "grad_norm": 29.683332443237305, + "learning_rate": 4.6421171644912295e-05, + "loss": 0.7007, + "step": 1227 + }, + { + "epoch": 0.8613010696124846, + "grad_norm": 29.98341178894043, + "learning_rate": 4.6413692899635764e-05, + "loss": 0.6981, + "step": 1228 + }, + { + "epoch": 0.8620024548483255, + "grad_norm": 30.50311279296875, + "learning_rate": 4.6406206952019014e-05, + "loss": 0.7341, + "step": 1229 + }, + { + "epoch": 0.8627038400841662, + "grad_norm": 1.5321385860443115, + "learning_rate": 4.6398713804579896e-05, + "loss": 0.0171, + "step": 1230 + }, + { + "epoch": 0.863405225320007, + "grad_norm": 25.173595428466797, + "learning_rate": 4.6391213459838676e-05, + "loss": 0.5393, + "step": 1231 + }, + { + "epoch": 0.8641066105558478, + "grad_norm": 1.9056795835494995, + "learning_rate": 4.638370592031804e-05, + "loss": 0.0216, + "step": 1232 + }, + { + "epoch": 0.8648079957916885, + "grad_norm": 54.3871955871582, + "learning_rate": 4.63761911885431e-05, + "loss": 1.1726, + "step": 1233 + }, + { + "epoch": 0.8655093810275294, + "grad_norm": 2.3975331783294678, + "learning_rate": 4.6368669267041364e-05, + "loss": 0.0284, + "step": 1234 + }, + { + "epoch": 0.8662107662633701, + "grad_norm": 2.3505735397338867, + "learning_rate": 4.636114015834279e-05, + "loss": 0.027, + "step": 1235 + }, + { + "epoch": 0.866912151499211, + "grad_norm": 19.353313446044922, + "learning_rate": 4.635360386497973e-05, + "loss": 0.4174, + "step": 1236 + }, + { + "epoch": 0.8676135367350517, + "grad_norm": 2.5313189029693604, + "learning_rate": 4.6346060389486976e-05, + "loss": 0.0301, + "step": 1237 + }, + { + "epoch": 0.8683149219708926, + "grad_norm": 2.249114751815796, + "learning_rate": 4.633850973440171e-05, + "loss": 0.0262, + "step": 1238 + }, + { + "epoch": 0.8690163072067333, + "grad_norm": 1.087750792503357, + "learning_rate": 4.633095190226355e-05, + "loss": 0.0116, + "step": 1239 + }, + { + "epoch": 0.8697176924425741, + "grad_norm": 0.7214330434799194, + "learning_rate": 4.63233868956145e-05, + "loss": 0.0075, + "step": 1240 + }, + { + "epoch": 0.8704190776784149, + "grad_norm": 0.5183669924736023, + "learning_rate": 4.6315814716999003e-05, + "loss": 0.0051, + "step": 1241 + }, + { + "epoch": 0.8711204629142556, + "grad_norm": 0.26819220185279846, + "learning_rate": 4.630823536896392e-05, + "loss": 0.0026, + "step": 1242 + }, + { + "epoch": 0.8718218481500964, + "grad_norm": 42.74302291870117, + "learning_rate": 4.6300648854058496e-05, + "loss": 1.4745, + "step": 1243 + }, + { + "epoch": 0.8725232333859372, + "grad_norm": 0.06916595250368118, + "learning_rate": 4.62930551748344e-05, + "loss": 0.0006, + "step": 1244 + }, + { + "epoch": 0.873224618621778, + "grad_norm": 0.04199717566370964, + "learning_rate": 4.6285454333845714e-05, + "loss": 0.0004, + "step": 1245 + }, + { + "epoch": 0.8739260038576188, + "grad_norm": 46.651554107666016, + "learning_rate": 4.6277846333648926e-05, + "loss": 2.1377, + "step": 1246 + }, + { + "epoch": 0.8746273890934596, + "grad_norm": 0.019042061641812325, + "learning_rate": 4.6270231176802936e-05, + "loss": 0.0002, + "step": 1247 + }, + { + "epoch": 0.8753287743293003, + "grad_norm": 43.70697784423828, + "learning_rate": 4.626260886586903e-05, + "loss": 1.9754, + "step": 1248 + }, + { + "epoch": 0.8760301595651412, + "grad_norm": 0.03968697413802147, + "learning_rate": 4.625497940341094e-05, + "loss": 0.0003, + "step": 1249 + }, + { + "epoch": 0.8767315448009819, + "grad_norm": 0.023654410615563393, + "learning_rate": 4.624734279199476e-05, + "loss": 0.0002, + "step": 1250 + }, + { + "epoch": 0.8774329300368228, + "grad_norm": 0.05060436204075813, + "learning_rate": 4.623969903418903e-05, + "loss": 0.0004, + "step": 1251 + }, + { + "epoch": 0.8781343152726635, + "grad_norm": 0.062442123889923096, + "learning_rate": 4.623204813256465e-05, + "loss": 0.0006, + "step": 1252 + }, + { + "epoch": 0.8788357005085043, + "grad_norm": 0.05655084550380707, + "learning_rate": 4.622439008969497e-05, + "loss": 0.0005, + "step": 1253 + }, + { + "epoch": 0.8795370857443451, + "grad_norm": 0.05976495519280434, + "learning_rate": 4.621672490815569e-05, + "loss": 0.0005, + "step": 1254 + }, + { + "epoch": 0.8802384709801858, + "grad_norm": 0.06660816818475723, + "learning_rate": 4.620905259052496e-05, + "loss": 0.0006, + "step": 1255 + }, + { + "epoch": 0.8809398562160267, + "grad_norm": 45.397789001464844, + "learning_rate": 4.6201373139383284e-05, + "loss": 1.7976, + "step": 1256 + }, + { + "epoch": 0.8816412414518674, + "grad_norm": 44.1479606628418, + "learning_rate": 4.619368655731362e-05, + "loss": 1.7646, + "step": 1257 + }, + { + "epoch": 0.8823426266877082, + "grad_norm": 0.13133244216442108, + "learning_rate": 4.618599284690127e-05, + "loss": 0.0012, + "step": 1258 + }, + { + "epoch": 0.883044011923549, + "grad_norm": 82.86041259765625, + "learning_rate": 4.6178292010733984e-05, + "loss": 2.727, + "step": 1259 + }, + { + "epoch": 0.8837453971593898, + "grad_norm": 36.786067962646484, + "learning_rate": 4.6170584051401846e-05, + "loss": 1.0486, + "step": 1260 + }, + { + "epoch": 0.8844467823952306, + "grad_norm": 43.94260787963867, + "learning_rate": 4.6162868971497396e-05, + "loss": 1.3111, + "step": 1261 + }, + { + "epoch": 0.8851481676310714, + "grad_norm": 85.85136413574219, + "learning_rate": 4.615514677361554e-05, + "loss": 2.2235, + "step": 1262 + }, + { + "epoch": 0.8858495528669121, + "grad_norm": 35.138553619384766, + "learning_rate": 4.6147417460353596e-05, + "loss": 0.6498, + "step": 1263 + }, + { + "epoch": 0.886550938102753, + "grad_norm": 5.97971773147583, + "learning_rate": 4.613968103431124e-05, + "loss": 0.0857, + "step": 1264 + }, + { + "epoch": 0.8872523233385937, + "grad_norm": 7.489769458770752, + "learning_rate": 4.6131937498090575e-05, + "loss": 0.1154, + "step": 1265 + }, + { + "epoch": 0.8879537085744345, + "grad_norm": 9.91996955871582, + "learning_rate": 4.612418685429608e-05, + "loss": 0.1685, + "step": 1266 + }, + { + "epoch": 0.8886550938102753, + "grad_norm": 2.7866861820220947, + "learning_rate": 4.611642910553463e-05, + "loss": 0.2187, + "step": 1267 + }, + { + "epoch": 0.889356479046116, + "grad_norm": 9.682534217834473, + "learning_rate": 4.610866425441549e-05, + "loss": 0.1669, + "step": 1268 + }, + { + "epoch": 0.8900578642819569, + "grad_norm": 8.847615242004395, + "learning_rate": 4.6100892303550305e-05, + "loss": 0.1458, + "step": 1269 + }, + { + "epoch": 0.8907592495177976, + "grad_norm": 6.194153308868408, + "learning_rate": 4.609311325555311e-05, + "loss": 0.0885, + "step": 1270 + }, + { + "epoch": 0.8914606347536385, + "grad_norm": 11.383880615234375, + "learning_rate": 4.6085327113040346e-05, + "loss": 0.2945, + "step": 1271 + }, + { + "epoch": 0.8921620199894792, + "grad_norm": 41.971961975097656, + "learning_rate": 4.607753387863082e-05, + "loss": 0.8504, + "step": 1272 + }, + { + "epoch": 0.89286340522532, + "grad_norm": 15.673029899597168, + "learning_rate": 4.6069733554945725e-05, + "loss": 0.3576, + "step": 1273 + }, + { + "epoch": 0.8935647904611608, + "grad_norm": 2.994936466217041, + "learning_rate": 4.606192614460865e-05, + "loss": 0.0371, + "step": 1274 + }, + { + "epoch": 0.8942661756970016, + "grad_norm": 15.890610694885254, + "learning_rate": 4.6054111650245544e-05, + "loss": 0.3475, + "step": 1275 + }, + { + "epoch": 0.8949675609328424, + "grad_norm": 3.3848869800567627, + "learning_rate": 4.604629007448479e-05, + "loss": 0.0433, + "step": 1276 + }, + { + "epoch": 0.8956689461686832, + "grad_norm": 26.11835289001465, + "learning_rate": 4.603846141995708e-05, + "loss": 0.5389, + "step": 1277 + }, + { + "epoch": 0.8963703314045239, + "grad_norm": 3.1392459869384766, + "learning_rate": 4.6030625689295546e-05, + "loss": 0.0383, + "step": 1278 + }, + { + "epoch": 0.8970717166403647, + "grad_norm": 3.49643611907959, + "learning_rate": 4.6022782885135674e-05, + "loss": 0.0445, + "step": 1279 + }, + { + "epoch": 0.8977731018762055, + "grad_norm": 2.069392204284668, + "learning_rate": 4.601493301011534e-05, + "loss": 0.0228, + "step": 1280 + }, + { + "epoch": 0.8984744871120462, + "grad_norm": 1.4139840602874756, + "learning_rate": 4.6007076066874786e-05, + "loss": 0.016, + "step": 1281 + }, + { + "epoch": 0.8991758723478871, + "grad_norm": 0.8678575158119202, + "learning_rate": 4.599921205805663e-05, + "loss": 0.0092, + "step": 1282 + }, + { + "epoch": 0.8998772575837278, + "grad_norm": 0.3188234567642212, + "learning_rate": 4.5991340986305896e-05, + "loss": 0.0031, + "step": 1283 + }, + { + "epoch": 0.9005786428195687, + "grad_norm": 0.13479161262512207, + "learning_rate": 4.598346285426994e-05, + "loss": 0.0012, + "step": 1284 + }, + { + "epoch": 0.9012800280554094, + "grad_norm": 44.85018539428711, + "learning_rate": 4.5975577664598535e-05, + "loss": 1.8893, + "step": 1285 + }, + { + "epoch": 0.9019814132912503, + "grad_norm": 0.030871881172060966, + "learning_rate": 4.596768541994379e-05, + "loss": 0.0003, + "step": 1286 + }, + { + "epoch": 0.902682798527091, + "grad_norm": 0.025855490937829018, + "learning_rate": 4.595978612296021e-05, + "loss": 0.0002, + "step": 1287 + }, + { + "epoch": 0.9033841837629318, + "grad_norm": 46.28804397583008, + "learning_rate": 4.595187977630467e-05, + "loss": 2.4086, + "step": 1288 + }, + { + "epoch": 0.9040855689987726, + "grad_norm": 46.17903518676758, + "learning_rate": 4.5943966382636413e-05, + "loss": 2.5441, + "step": 1289 + }, + { + "epoch": 0.9047869542346134, + "grad_norm": 48.56827926635742, + "learning_rate": 4.593604594461704e-05, + "loss": 2.5474, + "step": 1290 + }, + { + "epoch": 0.9054883394704542, + "grad_norm": 47.86493682861328, + "learning_rate": 4.592811846491055e-05, + "loss": 2.1781, + "step": 1291 + }, + { + "epoch": 0.9061897247062949, + "grad_norm": 43.444766998291016, + "learning_rate": 4.592018394618328e-05, + "loss": 1.8371, + "step": 1292 + }, + { + "epoch": 0.9068911099421357, + "grad_norm": 46.283565521240234, + "learning_rate": 4.5912242391103944e-05, + "loss": 1.5648, + "step": 1293 + }, + { + "epoch": 0.9075924951779765, + "grad_norm": 0.2855938971042633, + "learning_rate": 4.5904293802343646e-05, + "loss": 0.0028, + "step": 1294 + }, + { + "epoch": 0.9082938804138173, + "grad_norm": 44.066627502441406, + "learning_rate": 4.589633818257582e-05, + "loss": 1.3621, + "step": 1295 + }, + { + "epoch": 0.908995265649658, + "grad_norm": 1.1252837181091309, + "learning_rate": 4.588837553447628e-05, + "loss": 0.0122, + "step": 1296 + }, + { + "epoch": 0.9096966508854989, + "grad_norm": 1.7556012868881226, + "learning_rate": 4.588040586072321e-05, + "loss": 0.0199, + "step": 1297 + }, + { + "epoch": 0.9103980361213396, + "grad_norm": 2.0839264392852783, + "learning_rate": 4.5872429163997163e-05, + "loss": 0.0248, + "step": 1298 + }, + { + "epoch": 0.9110994213571805, + "grad_norm": 2.153773784637451, + "learning_rate": 4.5864445446981025e-05, + "loss": 0.0254, + "step": 1299 + }, + { + "epoch": 0.9118008065930212, + "grad_norm": 2.1063156127929688, + "learning_rate": 4.5856454712360067e-05, + "loss": 0.0235, + "step": 1300 + }, + { + "epoch": 0.912502191828862, + "grad_norm": 1.26010000705719, + "learning_rate": 4.584845696282192e-05, + "loss": 0.0134, + "step": 1301 + }, + { + "epoch": 0.9132035770647028, + "grad_norm": 0.7219848036766052, + "learning_rate": 4.584045220105656e-05, + "loss": 0.0075, + "step": 1302 + }, + { + "epoch": 0.9139049623005435, + "grad_norm": 0.3715592324733734, + "learning_rate": 4.583244042975634e-05, + "loss": 0.0037, + "step": 1303 + }, + { + "epoch": 0.9146063475363844, + "grad_norm": 0.16978369653224945, + "learning_rate": 4.582442165161596e-05, + "loss": 0.0017, + "step": 1304 + }, + { + "epoch": 0.9153077327722251, + "grad_norm": 0.10813824832439423, + "learning_rate": 4.5816395869332476e-05, + "loss": 0.001, + "step": 1305 + }, + { + "epoch": 0.916009118008066, + "grad_norm": 0.05026814341545105, + "learning_rate": 4.5808363085605296e-05, + "loss": 0.0005, + "step": 1306 + }, + { + "epoch": 0.9167105032439067, + "grad_norm": 0.025502556934952736, + "learning_rate": 4.58003233031362e-05, + "loss": 0.0002, + "step": 1307 + }, + { + "epoch": 0.9174118884797475, + "grad_norm": 45.29407501220703, + "learning_rate": 4.579227652462928e-05, + "loss": 2.3544, + "step": 1308 + }, + { + "epoch": 0.9181132737155883, + "grad_norm": 0.012745014391839504, + "learning_rate": 4.578422275279105e-05, + "loss": 0.0001, + "step": 1309 + }, + { + "epoch": 0.9188146589514291, + "grad_norm": 0.009426114149391651, + "learning_rate": 4.577616199033033e-05, + "loss": 0.0001, + "step": 1310 + }, + { + "epoch": 0.9195160441872698, + "grad_norm": 0.008077832870185375, + "learning_rate": 4.5768094239958285e-05, + "loss": 0.0001, + "step": 1311 + }, + { + "epoch": 0.9202174294231107, + "grad_norm": 0.006128115579485893, + "learning_rate": 4.576001950438845e-05, + "loss": 0.0001, + "step": 1312 + }, + { + "epoch": 0.9209188146589514, + "grad_norm": 88.39704895019531, + "learning_rate": 4.575193778633671e-05, + "loss": 5.5453, + "step": 1313 + }, + { + "epoch": 0.9216201998947923, + "grad_norm": 45.73112869262695, + "learning_rate": 4.574384908852128e-05, + "loss": 2.724, + "step": 1314 + }, + { + "epoch": 0.922321585130633, + "grad_norm": 45.08832550048828, + "learning_rate": 4.573575341366274e-05, + "loss": 2.5745, + "step": 1315 + }, + { + "epoch": 0.9230229703664737, + "grad_norm": 0.01642024703323841, + "learning_rate": 4.572765076448402e-05, + "loss": 0.0001, + "step": 1316 + }, + { + "epoch": 0.9237243556023146, + "grad_norm": 0.0225964467972517, + "learning_rate": 4.5719541143710375e-05, + "loss": 0.0002, + "step": 1317 + }, + { + "epoch": 0.9244257408381553, + "grad_norm": 0.03809621185064316, + "learning_rate": 4.571142455406943e-05, + "loss": 0.0003, + "step": 1318 + }, + { + "epoch": 0.9251271260739962, + "grad_norm": 42.60374450683594, + "learning_rate": 4.5703300998291124e-05, + "loss": 1.805, + "step": 1319 + }, + { + "epoch": 0.9258285113098369, + "grad_norm": 0.08730126917362213, + "learning_rate": 4.569517047910777e-05, + "loss": 0.0008, + "step": 1320 + }, + { + "epoch": 0.9265298965456777, + "grad_norm": 119.11152648925781, + "learning_rate": 4.5687032999254e-05, + "loss": 4.0878, + "step": 1321 + }, + { + "epoch": 0.9272312817815185, + "grad_norm": 0.3243617117404938, + "learning_rate": 4.567888856146681e-05, + "loss": 0.0031, + "step": 1322 + }, + { + "epoch": 0.9279326670173593, + "grad_norm": 0.4152386784553528, + "learning_rate": 4.567073716848551e-05, + "loss": 0.004, + "step": 1323 + }, + { + "epoch": 0.9286340522532001, + "grad_norm": 30.812305450439453, + "learning_rate": 4.566257882305176e-05, + "loss": 0.7747, + "step": 1324 + }, + { + "epoch": 0.9293354374890409, + "grad_norm": 30.550561904907227, + "learning_rate": 4.565441352790957e-05, + "loss": 0.7148, + "step": 1325 + }, + { + "epoch": 0.9300368227248816, + "grad_norm": 79.6876220703125, + "learning_rate": 4.564624128580528e-05, + "loss": 1.728, + "step": 1326 + }, + { + "epoch": 0.9307382079607224, + "grad_norm": 3.2627832889556885, + "learning_rate": 4.5638062099487545e-05, + "loss": 0.0427, + "step": 1327 + }, + { + "epoch": 0.9314395931965632, + "grad_norm": 10.727757453918457, + "learning_rate": 4.5629875971707395e-05, + "loss": 0.2789, + "step": 1328 + }, + { + "epoch": 0.932140978432404, + "grad_norm": 5.946654319763184, + "learning_rate": 4.562168290521817e-05, + "loss": 0.0887, + "step": 1329 + }, + { + "epoch": 0.9328423636682448, + "grad_norm": 1.3037784099578857, + "learning_rate": 4.561348290277554e-05, + "loss": 0.1909, + "step": 1330 + }, + { + "epoch": 0.9335437489040855, + "grad_norm": 8.197032928466797, + "learning_rate": 4.560527596713752e-05, + "loss": 0.1388, + "step": 1331 + }, + { + "epoch": 0.9342451341399264, + "grad_norm": 8.003156661987305, + "learning_rate": 4.559706210106446e-05, + "loss": 0.1299, + "step": 1332 + }, + { + "epoch": 0.9349465193757671, + "grad_norm": 13.647798538208008, + "learning_rate": 4.558884130731903e-05, + "loss": 0.3574, + "step": 1333 + }, + { + "epoch": 0.935647904611608, + "grad_norm": 5.755825996398926, + "learning_rate": 4.558061358866623e-05, + "loss": 0.0832, + "step": 1334 + }, + { + "epoch": 0.9363492898474487, + "grad_norm": 21.87740135192871, + "learning_rate": 4.55723789478734e-05, + "loss": 0.4666, + "step": 1335 + }, + { + "epoch": 0.9370506750832895, + "grad_norm": 9.851075172424316, + "learning_rate": 4.5564137387710196e-05, + "loss": 0.2624, + "step": 1336 + }, + { + "epoch": 0.9377520603191303, + "grad_norm": 4.983445644378662, + "learning_rate": 4.555588891094862e-05, + "loss": 0.0688, + "step": 1337 + }, + { + "epoch": 0.9384534455549711, + "grad_norm": 4.145101547241211, + "learning_rate": 4.554763352036296e-05, + "loss": 0.0537, + "step": 1338 + }, + { + "epoch": 0.9391548307908119, + "grad_norm": 4.348505020141602, + "learning_rate": 4.5539371218729894e-05, + "loss": 0.058, + "step": 1339 + }, + { + "epoch": 0.9398562160266526, + "grad_norm": 16.51106834411621, + "learning_rate": 4.5531102008828354e-05, + "loss": 0.3685, + "step": 1340 + }, + { + "epoch": 0.9405576012624934, + "grad_norm": 2.958770513534546, + "learning_rate": 4.5522825893439645e-05, + "loss": 0.0375, + "step": 1341 + }, + { + "epoch": 0.9412589864983342, + "grad_norm": 1.705260157585144, + "learning_rate": 4.5514542875347376e-05, + "loss": 0.0194, + "step": 1342 + }, + { + "epoch": 0.941960371734175, + "grad_norm": 28.18739891052246, + "learning_rate": 4.550625295733748e-05, + "loss": 0.6777, + "step": 1343 + }, + { + "epoch": 0.9426617569700158, + "grad_norm": 39.53862380981445, + "learning_rate": 4.5497956142198214e-05, + "loss": 1.1208, + "step": 1344 + }, + { + "epoch": 0.9433631422058566, + "grad_norm": 1.0771280527114868, + "learning_rate": 4.5489652432720145e-05, + "loss": 0.0118, + "step": 1345 + }, + { + "epoch": 0.9440645274416973, + "grad_norm": 96.40424346923828, + "learning_rate": 4.548134183169617e-05, + "loss": 2.4673, + "step": 1346 + }, + { + "epoch": 0.9447659126775382, + "grad_norm": 0.5970985293388367, + "learning_rate": 4.547302434192149e-05, + "loss": 0.0061, + "step": 1347 + }, + { + "epoch": 0.9454672979133789, + "grad_norm": 36.97736358642578, + "learning_rate": 4.546469996619364e-05, + "loss": 0.9705, + "step": 1348 + }, + { + "epoch": 0.9461686831492198, + "grad_norm": 0.6519697308540344, + "learning_rate": 4.545636870731247e-05, + "loss": 0.0066, + "step": 1349 + }, + { + "epoch": 0.9468700683850605, + "grad_norm": 0.4174421429634094, + "learning_rate": 4.544803056808013e-05, + "loss": 0.0042, + "step": 1350 + }, + { + "epoch": 0.9475714536209012, + "grad_norm": 0.4273722469806671, + "learning_rate": 4.5439685551301094e-05, + "loss": 0.0043, + "step": 1351 + }, + { + "epoch": 0.9482728388567421, + "grad_norm": 0.3083280324935913, + "learning_rate": 4.543133365978214e-05, + "loss": 0.003, + "step": 1352 + }, + { + "epoch": 0.9489742240925828, + "grad_norm": 0.2918597161769867, + "learning_rate": 4.542297489633238e-05, + "loss": 0.0028, + "step": 1353 + }, + { + "epoch": 0.9496756093284237, + "grad_norm": 47.360557556152344, + "learning_rate": 4.541460926376322e-05, + "loss": 1.7171, + "step": 1354 + }, + { + "epoch": 0.9503769945642644, + "grad_norm": 81.7466049194336, + "learning_rate": 4.5406236764888374e-05, + "loss": 2.8326, + "step": 1355 + }, + { + "epoch": 0.9510783798001052, + "grad_norm": 39.265384674072266, + "learning_rate": 4.539785740252388e-05, + "loss": 1.3337, + "step": 1356 + }, + { + "epoch": 0.951779765035946, + "grad_norm": 0.28257325291633606, + "learning_rate": 4.5389471179488074e-05, + "loss": 0.0027, + "step": 1357 + }, + { + "epoch": 0.9524811502717868, + "grad_norm": 41.259849548339844, + "learning_rate": 4.538107809860159e-05, + "loss": 1.2651, + "step": 1358 + }, + { + "epoch": 0.9531825355076275, + "grad_norm": 0.5483337640762329, + "learning_rate": 4.537267816268739e-05, + "loss": 0.0055, + "step": 1359 + }, + { + "epoch": 0.9538839207434684, + "grad_norm": 0.9649966359138489, + "learning_rate": 4.536427137457072e-05, + "loss": 0.0098, + "step": 1360 + }, + { + "epoch": 0.9545853059793091, + "grad_norm": 28.19986915588379, + "learning_rate": 4.5355857737079166e-05, + "loss": 0.6184, + "step": 1361 + }, + { + "epoch": 0.95528669121515, + "grad_norm": 1.310431957244873, + "learning_rate": 4.534743725304257e-05, + "loss": 0.0142, + "step": 1362 + }, + { + "epoch": 0.9559880764509907, + "grad_norm": 1.170506477355957, + "learning_rate": 4.5339009925293105e-05, + "loss": 0.0125, + "step": 1363 + }, + { + "epoch": 0.9566894616868314, + "grad_norm": 1.135798454284668, + "learning_rate": 4.533057575666526e-05, + "loss": 0.0118, + "step": 1364 + }, + { + "epoch": 0.9573908469226723, + "grad_norm": 0.7438721060752869, + "learning_rate": 4.532213474999578e-05, + "loss": 0.0077, + "step": 1365 + }, + { + "epoch": 0.958092232158513, + "grad_norm": 0.46775007247924805, + "learning_rate": 4.531368690812376e-05, + "loss": 0.0046, + "step": 1366 + }, + { + "epoch": 0.9587936173943539, + "grad_norm": 0.2829347848892212, + "learning_rate": 4.530523223389056e-05, + "loss": 0.0027, + "step": 1367 + }, + { + "epoch": 0.9594950026301946, + "grad_norm": 0.12059792876243591, + "learning_rate": 4.529677073013984e-05, + "loss": 0.0011, + "step": 1368 + }, + { + "epoch": 0.9601963878660355, + "grad_norm": 41.76416015625, + "learning_rate": 4.5288302399717575e-05, + "loss": 1.5052, + "step": 1369 + }, + { + "epoch": 0.9608977731018762, + "grad_norm": 0.05551628768444061, + "learning_rate": 4.5279827245472025e-05, + "loss": 0.0005, + "step": 1370 + }, + { + "epoch": 0.961599158337717, + "grad_norm": 44.579280853271484, + "learning_rate": 4.527134527025374e-05, + "loss": 1.9268, + "step": 1371 + }, + { + "epoch": 0.9623005435735578, + "grad_norm": 92.77808380126953, + "learning_rate": 4.526285647691559e-05, + "loss": 3.8084, + "step": 1372 + }, + { + "epoch": 0.9630019288093986, + "grad_norm": 0.06515678018331528, + "learning_rate": 4.5254360868312694e-05, + "loss": 0.0006, + "step": 1373 + }, + { + "epoch": 0.9637033140452393, + "grad_norm": 44.31393814086914, + "learning_rate": 4.5245858447302506e-05, + "loss": 1.8936, + "step": 1374 + }, + { + "epoch": 0.9644046992810801, + "grad_norm": 42.123191833496094, + "learning_rate": 4.523734921674475e-05, + "loss": 1.5489, + "step": 1375 + }, + { + "epoch": 0.9651060845169209, + "grad_norm": 78.27152252197266, + "learning_rate": 4.522883317950144e-05, + "loss": 1.4444, + "step": 1376 + }, + { + "epoch": 0.9658074697527617, + "grad_norm": 0.4181966185569763, + "learning_rate": 4.5220310338436885e-05, + "loss": 0.0041, + "step": 1377 + }, + { + "epoch": 0.9665088549886025, + "grad_norm": 61.79771423339844, + "learning_rate": 4.521178069641768e-05, + "loss": 1.5033, + "step": 1378 + }, + { + "epoch": 0.9672102402244432, + "grad_norm": 1.026915431022644, + "learning_rate": 4.520324425631271e-05, + "loss": 0.011, + "step": 1379 + }, + { + "epoch": 0.9679116254602841, + "grad_norm": 1.1025463342666626, + "learning_rate": 4.519470102099315e-05, + "loss": 0.0117, + "step": 1380 + }, + { + "epoch": 0.9686130106961248, + "grad_norm": 1.5587780475616455, + "learning_rate": 4.518615099333243e-05, + "loss": 0.018, + "step": 1381 + }, + { + "epoch": 0.9693143959319657, + "grad_norm": 1.4015796184539795, + "learning_rate": 4.517759417620633e-05, + "loss": 0.016, + "step": 1382 + }, + { + "epoch": 0.9700157811678064, + "grad_norm": 1.1269611120224, + "learning_rate": 4.516903057249284e-05, + "loss": 0.0124, + "step": 1383 + }, + { + "epoch": 0.9707171664036472, + "grad_norm": 64.7952651977539, + "learning_rate": 4.516046018507229e-05, + "loss": 1.7053, + "step": 1384 + }, + { + "epoch": 0.971418551639488, + "grad_norm": 36.87698745727539, + "learning_rate": 4.515188301682725e-05, + "loss": 0.8805, + "step": 1385 + }, + { + "epoch": 0.9721199368753288, + "grad_norm": 39.0102653503418, + "learning_rate": 4.5143299070642584e-05, + "loss": 1.2212, + "step": 1386 + }, + { + "epoch": 0.9728213221111696, + "grad_norm": 62.35078811645508, + "learning_rate": 4.5134708349405457e-05, + "loss": 1.6357, + "step": 1387 + }, + { + "epoch": 0.9735227073470103, + "grad_norm": 31.161195755004883, + "learning_rate": 4.5126110856005285e-05, + "loss": 0.7379, + "step": 1388 + }, + { + "epoch": 0.9742240925828511, + "grad_norm": 1.4460724592208862, + "learning_rate": 4.511750659333379e-05, + "loss": 0.0166, + "step": 1389 + }, + { + "epoch": 0.9749254778186919, + "grad_norm": 1.6871577501296997, + "learning_rate": 4.5108895564284924e-05, + "loss": 0.0201, + "step": 1390 + }, + { + "epoch": 0.9756268630545327, + "grad_norm": 18.601655960083008, + "learning_rate": 4.510027777175497e-05, + "loss": 0.3773, + "step": 1391 + }, + { + "epoch": 0.9763282482903735, + "grad_norm": 24.591615676879883, + "learning_rate": 4.509165321864245e-05, + "loss": 0.6012, + "step": 1392 + }, + { + "epoch": 0.9770296335262143, + "grad_norm": 2.3374860286712646, + "learning_rate": 4.5083021907848164e-05, + "loss": 0.0286, + "step": 1393 + }, + { + "epoch": 0.977731018762055, + "grad_norm": 3.1211438179016113, + "learning_rate": 4.50743838422752e-05, + "loss": 0.0388, + "step": 1394 + }, + { + "epoch": 0.9784324039978959, + "grad_norm": 15.115840911865234, + "learning_rate": 4.50657390248289e-05, + "loss": 0.3451, + "step": 1395 + }, + { + "epoch": 0.9791337892337366, + "grad_norm": 13.081830978393555, + "learning_rate": 4.50570874584169e-05, + "loss": 0.3082, + "step": 1396 + }, + { + "epoch": 0.9798351744695775, + "grad_norm": 3.7496583461761475, + "learning_rate": 4.5048429145949076e-05, + "loss": 0.0503, + "step": 1397 + }, + { + "epoch": 0.9805365597054182, + "grad_norm": 22.880462646484375, + "learning_rate": 4.503976409033759e-05, + "loss": 0.4432, + "step": 1398 + }, + { + "epoch": 0.9812379449412589, + "grad_norm": 4.944215774536133, + "learning_rate": 4.503109229449688e-05, + "loss": 0.0706, + "step": 1399 + }, + { + "epoch": 0.9819393301770998, + "grad_norm": 4.719092845916748, + "learning_rate": 4.502241376134364e-05, + "loss": 0.0667, + "step": 1400 + }, + { + "epoch": 0.9826407154129405, + "grad_norm": 4.378533363342285, + "learning_rate": 4.5013728493796826e-05, + "loss": 0.0611, + "step": 1401 + }, + { + "epoch": 0.9833421006487814, + "grad_norm": 4.234534740447998, + "learning_rate": 4.5005036494777666e-05, + "loss": 0.055, + "step": 1402 + }, + { + "epoch": 0.9840434858846221, + "grad_norm": 2.942084789276123, + "learning_rate": 4.4996337767209656e-05, + "loss": 0.0382, + "step": 1403 + }, + { + "epoch": 0.9847448711204629, + "grad_norm": 1.5875838994979858, + "learning_rate": 4.498763231401855e-05, + "loss": 0.018, + "step": 1404 + }, + { + "epoch": 0.9854462563563037, + "grad_norm": 53.9989128112793, + "learning_rate": 4.497892013813235e-05, + "loss": 1.2552, + "step": 1405 + }, + { + "epoch": 0.9861476415921445, + "grad_norm": 28.49068832397461, + "learning_rate": 4.497020124248135e-05, + "loss": 0.7336, + "step": 1406 + }, + { + "epoch": 0.9868490268279853, + "grad_norm": 0.4417467713356018, + "learning_rate": 4.496147562999807e-05, + "loss": 0.0045, + "step": 1407 + }, + { + "epoch": 0.9875504120638261, + "grad_norm": 104.99906921386719, + "learning_rate": 4.4952743303617325e-05, + "loss": 3.0756, + "step": 1408 + }, + { + "epoch": 0.9882517972996668, + "grad_norm": 0.3293454647064209, + "learning_rate": 4.4944004266276164e-05, + "loss": 0.0033, + "step": 1409 + }, + { + "epoch": 0.9889531825355077, + "grad_norm": 41.633243560791016, + "learning_rate": 4.4935258520913894e-05, + "loss": 1.3174, + "step": 1410 + }, + { + "epoch": 0.9896545677713484, + "grad_norm": 0.23380160331726074, + "learning_rate": 4.4926506070472085e-05, + "loss": 0.0023, + "step": 1411 + }, + { + "epoch": 0.9903559530071891, + "grad_norm": 0.44888490438461304, + "learning_rate": 4.4917746917894556e-05, + "loss": 0.0047, + "step": 1412 + }, + { + "epoch": 0.99105733824303, + "grad_norm": 33.958248138427734, + "learning_rate": 4.490898106612739e-05, + "loss": 0.9667, + "step": 1413 + }, + { + "epoch": 0.9917587234788707, + "grad_norm": 0.3081722855567932, + "learning_rate": 4.490020851811892e-05, + "loss": 0.003, + "step": 1414 + }, + { + "epoch": 0.9924601087147116, + "grad_norm": 32.04133224487305, + "learning_rate": 4.4891429276819726e-05, + "loss": 0.9003, + "step": 1415 + }, + { + "epoch": 0.9931614939505523, + "grad_norm": 43.6740608215332, + "learning_rate": 4.488264334518264e-05, + "loss": 0.9288, + "step": 1416 + }, + { + "epoch": 0.9938628791863932, + "grad_norm": 29.450580596923828, + "learning_rate": 4.487385072616275e-05, + "loss": 0.7131, + "step": 1417 + }, + { + "epoch": 0.9945642644222339, + "grad_norm": 1.1718950271606445, + "learning_rate": 4.486505142271739e-05, + "loss": 0.0137, + "step": 1418 + }, + { + "epoch": 0.9952656496580747, + "grad_norm": 1.5927923917770386, + "learning_rate": 4.485624543780613e-05, + "loss": 0.0192, + "step": 1419 + }, + { + "epoch": 0.9959670348939155, + "grad_norm": 24.835670471191406, + "learning_rate": 4.4847432774390807e-05, + "loss": 0.5717, + "step": 1420 + }, + { + "epoch": 0.9966684201297563, + "grad_norm": 1.8400554656982422, + "learning_rate": 4.4838613435435505e-05, + "loss": 0.022, + "step": 1421 + }, + { + "epoch": 0.997369805365597, + "grad_norm": 2.3654353618621826, + "learning_rate": 4.482978742390654e-05, + "loss": 0.0298, + "step": 1422 + }, + { + "epoch": 0.9980711906014378, + "grad_norm": 24.44013023376465, + "learning_rate": 4.482095474277247e-05, + "loss": 0.5608, + "step": 1423 + }, + { + "epoch": 0.9987725758372786, + "grad_norm": 2.5648183822631836, + "learning_rate": 4.48121153950041e-05, + "loss": 0.034, + "step": 1424 + }, + { + "epoch": 0.9994739610731194, + "grad_norm": 2.0194385051727295, + "learning_rate": 4.4803269383574507e-05, + "loss": 0.0249, + "step": 1425 + } + ], + "logging_steps": 1, + "max_steps": 1425, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 4.863941975950848e+17, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}