|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9994739610731194, |
|
"eval_steps": 500, |
|
"global_step": 1425, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0007013852358407856, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 0.9295, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0014027704716815712, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 1.432, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0021041557075223566, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 0.3822, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0028055409433631424, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 0.6542, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.003506926179203928, |
|
"grad_norm": 29.30962562561035, |
|
"learning_rate": 1.7543859649122808e-07, |
|
"loss": 0.3371, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.004208311415044713, |
|
"grad_norm": 34.5477294921875, |
|
"learning_rate": 3.5087719298245616e-07, |
|
"loss": 0.1946, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.004909696650885499, |
|
"grad_norm": 55.77326965332031, |
|
"learning_rate": 5.263157894736843e-07, |
|
"loss": 0.5682, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.005611081886726285, |
|
"grad_norm": 52.84772491455078, |
|
"learning_rate": 7.017543859649123e-07, |
|
"loss": 0.3959, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.00631246712256707, |
|
"grad_norm": 44.6428337097168, |
|
"learning_rate": 8.771929824561404e-07, |
|
"loss": 0.2899, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.007013852358407856, |
|
"grad_norm": 69.89168548583984, |
|
"learning_rate": 1.0526315789473685e-06, |
|
"loss": 0.6255, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.007715237594248641, |
|
"grad_norm": NaN, |
|
"learning_rate": 1.0526315789473685e-06, |
|
"loss": 0.4706, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.008416622830089426, |
|
"grad_norm": 65.1708755493164, |
|
"learning_rate": 1.2280701754385965e-06, |
|
"loss": 0.6875, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.009118008065930211, |
|
"grad_norm": 62.44850158691406, |
|
"learning_rate": 1.4035087719298246e-06, |
|
"loss": 0.8691, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.009819393301770998, |
|
"grad_norm": 50.52750778198242, |
|
"learning_rate": 1.5789473684210528e-06, |
|
"loss": 0.5682, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.010520778537611783, |
|
"grad_norm": 38.50554656982422, |
|
"learning_rate": 1.7543859649122807e-06, |
|
"loss": 0.4118, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.01122216377345257, |
|
"grad_norm": Infinity, |
|
"learning_rate": 1.7543859649122807e-06, |
|
"loss": 0.4622, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.011923549009293355, |
|
"grad_norm": 41.367347717285156, |
|
"learning_rate": 1.929824561403509e-06, |
|
"loss": 0.2331, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.01262493424513414, |
|
"grad_norm": 47.666603088378906, |
|
"learning_rate": 2.105263157894737e-06, |
|
"loss": 0.489, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.013326319480974926, |
|
"grad_norm": NaN, |
|
"learning_rate": 2.105263157894737e-06, |
|
"loss": 6.2172, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.014027704716815711, |
|
"grad_norm": 161.6815643310547, |
|
"learning_rate": 2.2807017543859652e-06, |
|
"loss": 1.9947, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.014729089952656496, |
|
"grad_norm": 14.031229019165039, |
|
"learning_rate": 2.456140350877193e-06, |
|
"loss": 0.0529, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.015430475188497283, |
|
"grad_norm": 51.389862060546875, |
|
"learning_rate": 2.631578947368421e-06, |
|
"loss": 0.3272, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.016131860424338066, |
|
"grad_norm": 16.31728744506836, |
|
"learning_rate": 2.8070175438596493e-06, |
|
"loss": 0.0668, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.016833245660178853, |
|
"grad_norm": 32.19120407104492, |
|
"learning_rate": 2.9824561403508774e-06, |
|
"loss": 0.1298, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.01753463089601964, |
|
"grad_norm": 8.127516746520996, |
|
"learning_rate": 3.1578947368421056e-06, |
|
"loss": 0.0245, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.018236016131860423, |
|
"grad_norm": 31.48802375793457, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 0.1808, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.01893740136770121, |
|
"grad_norm": 174.7755584716797, |
|
"learning_rate": 3.5087719298245615e-06, |
|
"loss": 1.5976, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.019638786603541996, |
|
"grad_norm": 173.22021484375, |
|
"learning_rate": 3.6842105263157892e-06, |
|
"loss": 2.0731, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.020340171839382783, |
|
"grad_norm": 16.176433563232422, |
|
"learning_rate": 3.859649122807018e-06, |
|
"loss": 0.0533, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.021041557075223566, |
|
"grad_norm": 25.456438064575195, |
|
"learning_rate": 4.035087719298246e-06, |
|
"loss": 0.2239, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.021742942311064353, |
|
"grad_norm": 2.50178861618042, |
|
"learning_rate": 4.210526315789474e-06, |
|
"loss": 0.007, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.02244432754690514, |
|
"grad_norm": 8.369728088378906, |
|
"learning_rate": 4.3859649122807014e-06, |
|
"loss": 0.0312, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.023145712782745922, |
|
"grad_norm": 1.419299602508545, |
|
"learning_rate": 4.5614035087719304e-06, |
|
"loss": 0.0039, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.02384709801858671, |
|
"grad_norm": 21.602149963378906, |
|
"learning_rate": 4.736842105263159e-06, |
|
"loss": 0.1907, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.024548483254427496, |
|
"grad_norm": 96.2072525024414, |
|
"learning_rate": 4.912280701754386e-06, |
|
"loss": 0.4621, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.02524986849026828, |
|
"grad_norm": 186.85726928710938, |
|
"learning_rate": 5.087719298245614e-06, |
|
"loss": 3.6381, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.025951253726109066, |
|
"grad_norm": 222.0509033203125, |
|
"learning_rate": 5.263157894736842e-06, |
|
"loss": 2.1687, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.026652638961949852, |
|
"grad_norm": 395.810791015625, |
|
"learning_rate": 5.43859649122807e-06, |
|
"loss": 4.1469, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.027354024197790636, |
|
"grad_norm": 24.080060958862305, |
|
"learning_rate": 5.6140350877192985e-06, |
|
"loss": 0.0944, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.028055409433631422, |
|
"grad_norm": 23.396860122680664, |
|
"learning_rate": 5.789473684210527e-06, |
|
"loss": 0.2287, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.02875679466947221, |
|
"grad_norm": 26.664474487304688, |
|
"learning_rate": 5.964912280701755e-06, |
|
"loss": 0.2467, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.029458179905312992, |
|
"grad_norm": 0.3562469482421875, |
|
"learning_rate": 6.140350877192982e-06, |
|
"loss": 0.0005, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.03015956514115378, |
|
"grad_norm": Infinity, |
|
"learning_rate": 6.140350877192982e-06, |
|
"loss": 1.6967, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.030860950376994566, |
|
"grad_norm": 0.1253829002380371, |
|
"learning_rate": 6.315789473684211e-06, |
|
"loss": 0.0003, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.03156233561283535, |
|
"grad_norm": 18.994827270507812, |
|
"learning_rate": 6.4912280701754385e-06, |
|
"loss": 0.233, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.03226372084867613, |
|
"grad_norm": 34.921852111816406, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 0.3528, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.03296510608451692, |
|
"grad_norm": 202.69903564453125, |
|
"learning_rate": 6.842105263157896e-06, |
|
"loss": 2.7552, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.033666491320357705, |
|
"grad_norm": 359.0797424316406, |
|
"learning_rate": 7.017543859649123e-06, |
|
"loss": 2.112, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.03436787655619849, |
|
"grad_norm": 3.0262975692749023, |
|
"learning_rate": 7.192982456140351e-06, |
|
"loss": 0.0094, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.03506926179203928, |
|
"grad_norm": 504.8307189941406, |
|
"learning_rate": 7.3684210526315784e-06, |
|
"loss": 6.4525, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.03577064702788006, |
|
"grad_norm": 7.283820152282715, |
|
"learning_rate": 7.5438596491228074e-06, |
|
"loss": 0.0419, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.036472032263720845, |
|
"grad_norm": 10.193113327026367, |
|
"learning_rate": 7.719298245614036e-06, |
|
"loss": 0.0805, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.037173417499561635, |
|
"grad_norm": 21.247915267944336, |
|
"learning_rate": 7.894736842105263e-06, |
|
"loss": 0.1385, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.03787480273540242, |
|
"grad_norm": 12.750287055969238, |
|
"learning_rate": 8.070175438596492e-06, |
|
"loss": 0.1786, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.03857618797124321, |
|
"grad_norm": 186.3012237548828, |
|
"learning_rate": 8.245614035087721e-06, |
|
"loss": 2.5431, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.03927757320708399, |
|
"grad_norm": 0.6861621141433716, |
|
"learning_rate": 8.421052631578948e-06, |
|
"loss": 0.0019, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.039978958442924775, |
|
"grad_norm": 0.36288923025131226, |
|
"learning_rate": 8.596491228070176e-06, |
|
"loss": 0.0008, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.040680343678765565, |
|
"grad_norm": 0.08618122339248657, |
|
"learning_rate": 8.771929824561403e-06, |
|
"loss": 0.0001, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.04138172891460635, |
|
"grad_norm": 341.2626037597656, |
|
"learning_rate": 8.947368421052632e-06, |
|
"loss": 3.6742, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.04208311415044713, |
|
"grad_norm": 5.118091583251953, |
|
"learning_rate": 9.122807017543861e-06, |
|
"loss": 0.0128, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.04278449938628792, |
|
"grad_norm": 3.345242977142334, |
|
"learning_rate": 9.298245614035088e-06, |
|
"loss": 0.0095, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.043485884622128705, |
|
"grad_norm": 319.7175598144531, |
|
"learning_rate": 9.473684210526317e-06, |
|
"loss": 4.3047, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.04418726985796949, |
|
"grad_norm": 279.8736267089844, |
|
"learning_rate": 9.649122807017545e-06, |
|
"loss": 1.8524, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.04488865509381028, |
|
"grad_norm": 187.50588989257812, |
|
"learning_rate": 9.824561403508772e-06, |
|
"loss": 2.2166, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.04559004032965106, |
|
"grad_norm": 3.1667685508728027, |
|
"learning_rate": 1e-05, |
|
"loss": 0.0051, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.046291425565491845, |
|
"grad_norm": 347.2233581542969, |
|
"learning_rate": 1.0175438596491228e-05, |
|
"loss": 6.7317, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.046992810801332635, |
|
"grad_norm": 0.21984274685382843, |
|
"learning_rate": 1.0350877192982457e-05, |
|
"loss": 0.0006, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.04769419603717342, |
|
"grad_norm": 657.2495727539062, |
|
"learning_rate": 1.0526315789473684e-05, |
|
"loss": 12.4963, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.0483955812730142, |
|
"grad_norm": 180.615234375, |
|
"learning_rate": 1.0701754385964913e-05, |
|
"loss": 3.2259, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.04909696650885499, |
|
"grad_norm": 0.4523675739765167, |
|
"learning_rate": 1.087719298245614e-05, |
|
"loss": 0.0011, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.049798351744695775, |
|
"grad_norm": 241.3996124267578, |
|
"learning_rate": 1.1052631578947368e-05, |
|
"loss": 2.9636, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.05049973698053656, |
|
"grad_norm": 0.21977363526821136, |
|
"learning_rate": 1.1228070175438597e-05, |
|
"loss": 0.0005, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.05120112221637735, |
|
"grad_norm": 102.25936889648438, |
|
"learning_rate": 1.1403508771929824e-05, |
|
"loss": 0.4978, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.05190250745221813, |
|
"grad_norm": 24.031539916992188, |
|
"learning_rate": 1.1578947368421053e-05, |
|
"loss": 0.119, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.052603892688058915, |
|
"grad_norm": 187.41812133789062, |
|
"learning_rate": 1.1754385964912282e-05, |
|
"loss": 1.0223, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.053305277923899705, |
|
"grad_norm": 45.128421783447266, |
|
"learning_rate": 1.192982456140351e-05, |
|
"loss": 0.4026, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.05400666315974049, |
|
"grad_norm": 12.320211410522461, |
|
"learning_rate": 1.2105263157894737e-05, |
|
"loss": 0.1124, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.05470804839558127, |
|
"grad_norm": 26.743837356567383, |
|
"learning_rate": 1.2280701754385964e-05, |
|
"loss": 0.1488, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.05540943363142206, |
|
"grad_norm": 30.013471603393555, |
|
"learning_rate": 1.2456140350877193e-05, |
|
"loss": 0.3796, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.056110818867262845, |
|
"grad_norm": 38.910221099853516, |
|
"learning_rate": 1.2631578947368422e-05, |
|
"loss": 0.7097, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.05681220410310363, |
|
"grad_norm": 26.001049041748047, |
|
"learning_rate": 1.2807017543859651e-05, |
|
"loss": 0.1451, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.05751358933894442, |
|
"grad_norm": 12.577742576599121, |
|
"learning_rate": 1.2982456140350877e-05, |
|
"loss": 0.0548, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.0582149745747852, |
|
"grad_norm": 334.09478759765625, |
|
"learning_rate": 1.3157894736842106e-05, |
|
"loss": 8.8326, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.058916359810625984, |
|
"grad_norm": 33.031837463378906, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 0.2555, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.059617745046466775, |
|
"grad_norm": 47.51133346557617, |
|
"learning_rate": 1.3508771929824562e-05, |
|
"loss": 0.4288, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.06031913028230756, |
|
"grad_norm": 41.160430908203125, |
|
"learning_rate": 1.3684210526315791e-05, |
|
"loss": 0.3075, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.06102051551814834, |
|
"grad_norm": 9.073790550231934, |
|
"learning_rate": 1.3859649122807017e-05, |
|
"loss": 0.045, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.06172190075398913, |
|
"grad_norm": 352.0498352050781, |
|
"learning_rate": 1.4035087719298246e-05, |
|
"loss": 1.3396, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.062423285989829914, |
|
"grad_norm": 36.5557746887207, |
|
"learning_rate": 1.4210526315789475e-05, |
|
"loss": 0.1926, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.0631246712256707, |
|
"grad_norm": 6.991324424743652, |
|
"learning_rate": 1.4385964912280702e-05, |
|
"loss": 0.0341, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.06382605646151149, |
|
"grad_norm": 6.0483078956604, |
|
"learning_rate": 1.4561403508771931e-05, |
|
"loss": 0.0219, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.06452744169735226, |
|
"grad_norm": 4.510119438171387, |
|
"learning_rate": 1.4736842105263157e-05, |
|
"loss": 0.0161, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.06522882693319305, |
|
"grad_norm": 0.8209487199783325, |
|
"learning_rate": 1.4912280701754386e-05, |
|
"loss": 0.0022, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.06593021216903384, |
|
"grad_norm": 0.0308829378336668, |
|
"learning_rate": 1.5087719298245615e-05, |
|
"loss": 0.0001, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.06663159740487462, |
|
"grad_norm": 171.68765258789062, |
|
"learning_rate": 1.5263157894736842e-05, |
|
"loss": 4.8668, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.06733298264071541, |
|
"grad_norm": 0.0008907412411645055, |
|
"learning_rate": 1.543859649122807e-05, |
|
"loss": 0.0, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.0680343678765562, |
|
"grad_norm": 0.0062671443447470665, |
|
"learning_rate": 1.56140350877193e-05, |
|
"loss": 0.0, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.06873575311239698, |
|
"grad_norm": 0.02987469919025898, |
|
"learning_rate": 1.5789473684210526e-05, |
|
"loss": 0.0001, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.06943713834823777, |
|
"grad_norm": 0.00033108692150563, |
|
"learning_rate": 1.5964912280701755e-05, |
|
"loss": 0.0, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.07013852358407856, |
|
"grad_norm": 0.21173974871635437, |
|
"learning_rate": 1.6140350877192984e-05, |
|
"loss": 0.0003, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.07083990881991933, |
|
"grad_norm": 0.0003154289734084159, |
|
"learning_rate": 1.6315789473684213e-05, |
|
"loss": 0.0, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.07154129405576012, |
|
"grad_norm": 0.0001488685520598665, |
|
"learning_rate": 1.6491228070175442e-05, |
|
"loss": 0.0, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.07224267929160091, |
|
"grad_norm": 1.3430032595351804e-05, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 0.0, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.07294406452744169, |
|
"grad_norm": 0.011110284365713596, |
|
"learning_rate": 1.6842105263157896e-05, |
|
"loss": 0.0, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.07364544976328248, |
|
"grad_norm": 1.7930891772266477e-05, |
|
"learning_rate": 1.7017543859649125e-05, |
|
"loss": 0.0, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.07434683499912327, |
|
"grad_norm": 5.653702146446449e-07, |
|
"learning_rate": 1.719298245614035e-05, |
|
"loss": 0.0, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.07504822023496406, |
|
"grad_norm": 135.4245147705078, |
|
"learning_rate": 1.736842105263158e-05, |
|
"loss": 6.4792, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.07574960547080484, |
|
"grad_norm": 153.99264526367188, |
|
"learning_rate": 1.7543859649122806e-05, |
|
"loss": 6.485, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.07645099070664563, |
|
"grad_norm": 159.35337829589844, |
|
"learning_rate": 1.7719298245614035e-05, |
|
"loss": 2.4936, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.07715237594248642, |
|
"grad_norm": 0.00017455461784265935, |
|
"learning_rate": 1.7894736842105264e-05, |
|
"loss": 0.0, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.0778537611783272, |
|
"grad_norm": 146.1802978515625, |
|
"learning_rate": 1.8070175438596493e-05, |
|
"loss": 4.3088, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.07855514641416798, |
|
"grad_norm": 0.022132212296128273, |
|
"learning_rate": 1.8245614035087722e-05, |
|
"loss": 0.0, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.07925653165000877, |
|
"grad_norm": 160.43975830078125, |
|
"learning_rate": 1.8421052631578947e-05, |
|
"loss": 4.5829, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.07995791688584955, |
|
"grad_norm": 164.45138549804688, |
|
"learning_rate": 1.8596491228070176e-05, |
|
"loss": 4.1681, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.08065930212169034, |
|
"grad_norm": 0.02766432799398899, |
|
"learning_rate": 1.8771929824561405e-05, |
|
"loss": 0.0001, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.08136068735753113, |
|
"grad_norm": 163.59400939941406, |
|
"learning_rate": 1.8947368421052634e-05, |
|
"loss": 3.0252, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.08206207259337191, |
|
"grad_norm": 7.11077356338501, |
|
"learning_rate": 1.9122807017543863e-05, |
|
"loss": 0.0309, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.0827634578292127, |
|
"grad_norm": 15.318146705627441, |
|
"learning_rate": 1.929824561403509e-05, |
|
"loss": 0.0856, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.08346484306505349, |
|
"grad_norm": 9.1824951171875, |
|
"learning_rate": 1.9473684210526315e-05, |
|
"loss": 0.0416, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.08416622830089426, |
|
"grad_norm": 22.720653533935547, |
|
"learning_rate": 1.9649122807017544e-05, |
|
"loss": 0.1365, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.08486761353673505, |
|
"grad_norm": 26.200096130371094, |
|
"learning_rate": 1.9824561403508773e-05, |
|
"loss": 0.2579, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.08556899877257584, |
|
"grad_norm": 15.874689102172852, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0832, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.08627038400841662, |
|
"grad_norm": 11.818103790283203, |
|
"learning_rate": 2.0175438596491227e-05, |
|
"loss": 0.0541, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.08697176924425741, |
|
"grad_norm": 116.5608139038086, |
|
"learning_rate": 2.0350877192982456e-05, |
|
"loss": 0.9623, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.0876731544800982, |
|
"grad_norm": 10.346467018127441, |
|
"learning_rate": 2.0526315789473685e-05, |
|
"loss": 0.0562, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.08837453971593898, |
|
"grad_norm": 137.41136169433594, |
|
"learning_rate": 2.0701754385964914e-05, |
|
"loss": 1.1566, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.08907592495177977, |
|
"grad_norm": 2.9712822437286377, |
|
"learning_rate": 2.0877192982456143e-05, |
|
"loss": 0.0104, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.08977731018762056, |
|
"grad_norm": 18.723491668701172, |
|
"learning_rate": 2.105263157894737e-05, |
|
"loss": 0.0667, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.09047869542346133, |
|
"grad_norm": 156.25631713867188, |
|
"learning_rate": 2.1228070175438598e-05, |
|
"loss": 2.4023, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.09118008065930212, |
|
"grad_norm": 298.97552490234375, |
|
"learning_rate": 2.1403508771929827e-05, |
|
"loss": 6.4928, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.09188146589514291, |
|
"grad_norm": 87.54296112060547, |
|
"learning_rate": 2.1578947368421053e-05, |
|
"loss": 0.5526, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.09258285113098369, |
|
"grad_norm": 0.25955039262771606, |
|
"learning_rate": 2.175438596491228e-05, |
|
"loss": 0.0007, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.09328423636682448, |
|
"grad_norm": 123.57398986816406, |
|
"learning_rate": 2.1929824561403507e-05, |
|
"loss": 1.0402, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.09398562160266527, |
|
"grad_norm": 238.33229064941406, |
|
"learning_rate": 2.2105263157894736e-05, |
|
"loss": 2.4997, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.09468700683850605, |
|
"grad_norm": 13.356133460998535, |
|
"learning_rate": 2.2280701754385965e-05, |
|
"loss": 0.0549, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.09538839207434684, |
|
"grad_norm": 44.56560134887695, |
|
"learning_rate": 2.2456140350877194e-05, |
|
"loss": 0.4078, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.09608977731018763, |
|
"grad_norm": 25.394866943359375, |
|
"learning_rate": 2.2631578947368423e-05, |
|
"loss": 0.1681, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.0967911625460284, |
|
"grad_norm": 28.068824768066406, |
|
"learning_rate": 2.280701754385965e-05, |
|
"loss": 0.174, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.0974925477818692, |
|
"grad_norm": 27.76630401611328, |
|
"learning_rate": 2.2982456140350878e-05, |
|
"loss": 0.1515, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.09819393301770998, |
|
"grad_norm": 21.082311630249023, |
|
"learning_rate": 2.3157894736842107e-05, |
|
"loss": 0.1379, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.09889531825355076, |
|
"grad_norm": 16.697647094726562, |
|
"learning_rate": 2.3333333333333336e-05, |
|
"loss": 0.1066, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.09959670348939155, |
|
"grad_norm": 127.46953582763672, |
|
"learning_rate": 2.3508771929824565e-05, |
|
"loss": 1.3053, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.10029808872523234, |
|
"grad_norm": 6.2219648361206055, |
|
"learning_rate": 2.368421052631579e-05, |
|
"loss": 0.0323, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.10099947396107312, |
|
"grad_norm": 0.1309046596288681, |
|
"learning_rate": 2.385964912280702e-05, |
|
"loss": 0.0004, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.1017008591969139, |
|
"grad_norm": 149.50645446777344, |
|
"learning_rate": 2.4035087719298245e-05, |
|
"loss": 1.8353, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.1024022444327547, |
|
"grad_norm": 159.4170684814453, |
|
"learning_rate": 2.4210526315789474e-05, |
|
"loss": 3.2714, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.10310362966859547, |
|
"grad_norm": 0.038805797696113586, |
|
"learning_rate": 2.4385964912280703e-05, |
|
"loss": 0.0001, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.10380501490443626, |
|
"grad_norm": 301.7401123046875, |
|
"learning_rate": 2.456140350877193e-05, |
|
"loss": 5.5114, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.10450640014027705, |
|
"grad_norm": 160.01580810546875, |
|
"learning_rate": 2.4736842105263158e-05, |
|
"loss": 3.0398, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.10520778537611783, |
|
"grad_norm": 1.0593466758728027, |
|
"learning_rate": 2.4912280701754387e-05, |
|
"loss": 0.0026, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.10590917061195862, |
|
"grad_norm": 0.1588684320449829, |
|
"learning_rate": 2.5087719298245616e-05, |
|
"loss": 0.0004, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.10661055584779941, |
|
"grad_norm": 0.12831349670886993, |
|
"learning_rate": 2.5263157894736845e-05, |
|
"loss": 0.0004, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.10731194108364019, |
|
"grad_norm": 1.908389925956726, |
|
"learning_rate": 2.5438596491228074e-05, |
|
"loss": 0.0062, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.10801332631948098, |
|
"grad_norm": 0.9305810928344727, |
|
"learning_rate": 2.5614035087719303e-05, |
|
"loss": 0.0029, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.10871471155532177, |
|
"grad_norm": 149.55746459960938, |
|
"learning_rate": 2.578947368421053e-05, |
|
"loss": 2.1251, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.10941609679116254, |
|
"grad_norm": 83.3541030883789, |
|
"learning_rate": 2.5964912280701754e-05, |
|
"loss": 0.4841, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.11011748202700333, |
|
"grad_norm": 0.37118402123451233, |
|
"learning_rate": 2.6140350877192983e-05, |
|
"loss": 0.0009, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.11081886726284412, |
|
"grad_norm": 212.9443359375, |
|
"learning_rate": 2.6315789473684212e-05, |
|
"loss": 1.6639, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.1115202524986849, |
|
"grad_norm": 1.386982798576355, |
|
"learning_rate": 2.6491228070175438e-05, |
|
"loss": 0.0049, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.11222163773452569, |
|
"grad_norm": 151.23684692382812, |
|
"learning_rate": 2.6666666666666667e-05, |
|
"loss": 1.913, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.11292302297036648, |
|
"grad_norm": 276.9901123046875, |
|
"learning_rate": 2.6842105263157896e-05, |
|
"loss": 5.0838, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.11362440820620726, |
|
"grad_norm": 7.047855377197266, |
|
"learning_rate": 2.7017543859649125e-05, |
|
"loss": 0.0422, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.11432579344204805, |
|
"grad_norm": 5.565479755401611, |
|
"learning_rate": 2.7192982456140354e-05, |
|
"loss": 0.0252, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.11502717867788884, |
|
"grad_norm": 11.038246154785156, |
|
"learning_rate": 2.7368421052631583e-05, |
|
"loss": 0.064, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.11572856391372961, |
|
"grad_norm": 1.4180843830108643, |
|
"learning_rate": 2.754385964912281e-05, |
|
"loss": 0.0048, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.1164299491495704, |
|
"grad_norm": 131.0841064453125, |
|
"learning_rate": 2.7719298245614034e-05, |
|
"loss": 1.1797, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.11713133438541119, |
|
"grad_norm": 286.4391784667969, |
|
"learning_rate": 2.7894736842105263e-05, |
|
"loss": 4.2848, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.11783271962125197, |
|
"grad_norm": 0.354282945394516, |
|
"learning_rate": 2.8070175438596492e-05, |
|
"loss": 0.001, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.11853410485709276, |
|
"grad_norm": 148.43812561035156, |
|
"learning_rate": 2.824561403508772e-05, |
|
"loss": 2.3637, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.11923549009293355, |
|
"grad_norm": 3.4153621196746826, |
|
"learning_rate": 2.842105263157895e-05, |
|
"loss": 0.0121, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.11993687532877433, |
|
"grad_norm": 6.812062740325928, |
|
"learning_rate": 2.8596491228070175e-05, |
|
"loss": 0.038, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.12063826056461512, |
|
"grad_norm": 0.6381294131278992, |
|
"learning_rate": 2.8771929824561404e-05, |
|
"loss": 0.002, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.1213396458004559, |
|
"grad_norm": 0.05846899002790451, |
|
"learning_rate": 2.8947368421052634e-05, |
|
"loss": 0.0002, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.12204103103629668, |
|
"grad_norm": 140.1736297607422, |
|
"learning_rate": 2.9122807017543863e-05, |
|
"loss": 4.0868, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.12274241627213747, |
|
"grad_norm": 1.7416075468063354, |
|
"learning_rate": 2.929824561403509e-05, |
|
"loss": 0.0057, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.12344380150797826, |
|
"grad_norm": 0.35068169236183167, |
|
"learning_rate": 2.9473684210526314e-05, |
|
"loss": 0.0009, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.12414518674381904, |
|
"grad_norm": 135.71665954589844, |
|
"learning_rate": 2.9649122807017543e-05, |
|
"loss": 2.1025, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.12484657197965983, |
|
"grad_norm": 134.73190307617188, |
|
"learning_rate": 2.9824561403508772e-05, |
|
"loss": 1.8685, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.1255479572155006, |
|
"grad_norm": 266.5934143066406, |
|
"learning_rate": 3e-05, |
|
"loss": 5.1122, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.1262493424513414, |
|
"grad_norm": 383.29931640625, |
|
"learning_rate": 3.017543859649123e-05, |
|
"loss": 10.0944, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.12695072768718219, |
|
"grad_norm": 3.7992992401123047, |
|
"learning_rate": 3.035087719298246e-05, |
|
"loss": 0.0161, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.12765211292302298, |
|
"grad_norm": 206.0655975341797, |
|
"learning_rate": 3.0526315789473684e-05, |
|
"loss": 1.5303, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.12835349815886377, |
|
"grad_norm": 13.431243896484375, |
|
"learning_rate": 3.0701754385964913e-05, |
|
"loss": 0.0912, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.12905488339470453, |
|
"grad_norm": 16.69830894470215, |
|
"learning_rate": 3.087719298245614e-05, |
|
"loss": 0.0824, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.12975626863054532, |
|
"grad_norm": 13.488637924194336, |
|
"learning_rate": 3.105263157894737e-05, |
|
"loss": 0.0746, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.1304576538663861, |
|
"grad_norm": 10.065983772277832, |
|
"learning_rate": 3.12280701754386e-05, |
|
"loss": 0.0503, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.1311590391022269, |
|
"grad_norm": 123.50023651123047, |
|
"learning_rate": 3.140350877192982e-05, |
|
"loss": 1.2874, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.1318604243380677, |
|
"grad_norm": 1.7797540426254272, |
|
"learning_rate": 3.157894736842105e-05, |
|
"loss": 0.0067, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.13256180957390848, |
|
"grad_norm": 0.33675551414489746, |
|
"learning_rate": 3.175438596491228e-05, |
|
"loss": 0.0011, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.13326319480974924, |
|
"grad_norm": 118.4156723022461, |
|
"learning_rate": 3.192982456140351e-05, |
|
"loss": 1.2059, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.13396458004559003, |
|
"grad_norm": 136.65228271484375, |
|
"learning_rate": 3.210526315789474e-05, |
|
"loss": 1.7856, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.13466596528143082, |
|
"grad_norm": 145.7415313720703, |
|
"learning_rate": 3.228070175438597e-05, |
|
"loss": 2.8991, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.1353673505172716, |
|
"grad_norm": 262.7060241699219, |
|
"learning_rate": 3.24561403508772e-05, |
|
"loss": 3.4817, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.1360687357531124, |
|
"grad_norm": 0.25727707147598267, |
|
"learning_rate": 3.2631578947368426e-05, |
|
"loss": 0.0008, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.1367701209889532, |
|
"grad_norm": 1.0702918767929077, |
|
"learning_rate": 3.2807017543859655e-05, |
|
"loss": 0.0037, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.13747150622479395, |
|
"grad_norm": 1.5917607545852661, |
|
"learning_rate": 3.2982456140350884e-05, |
|
"loss": 0.0055, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.13817289146063474, |
|
"grad_norm": 355.18511962890625, |
|
"learning_rate": 3.3157894736842106e-05, |
|
"loss": 11.4677, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.13887427669647553, |
|
"grad_norm": 1.7947548627853394, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 0.0074, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.13957566193231633, |
|
"grad_norm": 1.5046889781951904, |
|
"learning_rate": 3.3508771929824564e-05, |
|
"loss": 0.0058, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.14027704716815712, |
|
"grad_norm": 0.2553917467594147, |
|
"learning_rate": 3.368421052631579e-05, |
|
"loss": 0.0008, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.1409784324039979, |
|
"grad_norm": 0.773563027381897, |
|
"learning_rate": 3.385964912280702e-05, |
|
"loss": 0.0027, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.14167981763983867, |
|
"grad_norm": 166.71820068359375, |
|
"learning_rate": 3.403508771929825e-05, |
|
"loss": 6.1496, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.14238120287567946, |
|
"grad_norm": 0.017217637971043587, |
|
"learning_rate": 3.421052631578947e-05, |
|
"loss": 0.0001, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.14308258811152025, |
|
"grad_norm": 0.0036761912051588297, |
|
"learning_rate": 3.43859649122807e-05, |
|
"loss": 0.0, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.14378397334736104, |
|
"grad_norm": 0.000679672637488693, |
|
"learning_rate": 3.456140350877193e-05, |
|
"loss": 0.0, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.14448535858320183, |
|
"grad_norm": 0.00011312704009469599, |
|
"learning_rate": 3.473684210526316e-05, |
|
"loss": 0.0, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.14518674381904262, |
|
"grad_norm": 0.00012077714200131595, |
|
"learning_rate": 3.491228070175438e-05, |
|
"loss": 0.0, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.14588812905488338, |
|
"grad_norm": 0.00012628763215616345, |
|
"learning_rate": 3.508771929824561e-05, |
|
"loss": 0.0, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.14658951429072417, |
|
"grad_norm": 1.7496255168225616e-05, |
|
"learning_rate": 3.526315789473684e-05, |
|
"loss": 0.0, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.14729089952656496, |
|
"grad_norm": 1.5739196896902286e-05, |
|
"learning_rate": 3.543859649122807e-05, |
|
"loss": 0.0, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.14799228476240575, |
|
"grad_norm": 0.08704076707363129, |
|
"learning_rate": 3.56140350877193e-05, |
|
"loss": 0.0001, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.14869366999824654, |
|
"grad_norm": 0.0016435593133792281, |
|
"learning_rate": 3.578947368421053e-05, |
|
"loss": 0.0, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.14939505523408733, |
|
"grad_norm": 1.5830031633377075, |
|
"learning_rate": 3.5964912280701756e-05, |
|
"loss": 0.005, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.15009644046992812, |
|
"grad_norm": 0.00016859408060554415, |
|
"learning_rate": 3.6140350877192985e-05, |
|
"loss": 0.0, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.15079782570576888, |
|
"grad_norm": 1.2645250535570085e-06, |
|
"learning_rate": 3.6315789473684214e-05, |
|
"loss": 0.0, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.15149921094160967, |
|
"grad_norm": 336.5249328613281, |
|
"learning_rate": 3.6491228070175443e-05, |
|
"loss": 4.1939, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.15220059617745046, |
|
"grad_norm": 9.010884127746976e-07, |
|
"learning_rate": 3.6666666666666666e-05, |
|
"loss": 0.0, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.15290198141329125, |
|
"grad_norm": 5.530741304937692e-07, |
|
"learning_rate": 3.6842105263157895e-05, |
|
"loss": 0.0, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.15360336664913204, |
|
"grad_norm": 222.7361297607422, |
|
"learning_rate": 3.7017543859649124e-05, |
|
"loss": 0.7712, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.15430475188497284, |
|
"grad_norm": 123.51704406738281, |
|
"learning_rate": 3.719298245614035e-05, |
|
"loss": 7.8703, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.1550061371208136, |
|
"grad_norm": 119.08999633789062, |
|
"learning_rate": 3.736842105263158e-05, |
|
"loss": 6.6427, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.1557075223566544, |
|
"grad_norm": 0.005513960495591164, |
|
"learning_rate": 3.754385964912281e-05, |
|
"loss": 0.0, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.15640890759249518, |
|
"grad_norm": 0.021047895774245262, |
|
"learning_rate": 3.771929824561404e-05, |
|
"loss": 0.0, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.15711029282833597, |
|
"grad_norm": 0.001549599110148847, |
|
"learning_rate": 3.789473684210527e-05, |
|
"loss": 0.0, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.15781167806417676, |
|
"grad_norm": 0.008736948482692242, |
|
"learning_rate": 3.80701754385965e-05, |
|
"loss": 0.0, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.15851306330001755, |
|
"grad_norm": 0.03641926124691963, |
|
"learning_rate": 3.824561403508773e-05, |
|
"loss": 0.0001, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.1592144485358583, |
|
"grad_norm": 0.026789208874106407, |
|
"learning_rate": 3.842105263157895e-05, |
|
"loss": 0.0001, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.1599158337716991, |
|
"grad_norm": 130.79087829589844, |
|
"learning_rate": 3.859649122807018e-05, |
|
"loss": 3.4204, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.1606172190075399, |
|
"grad_norm": 404.3518981933594, |
|
"learning_rate": 3.877192982456141e-05, |
|
"loss": 5.6741, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.16131860424338068, |
|
"grad_norm": 5.39788818359375, |
|
"learning_rate": 3.894736842105263e-05, |
|
"loss": 0.0251, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.16201998947922147, |
|
"grad_norm": 18.25702667236328, |
|
"learning_rate": 3.912280701754386e-05, |
|
"loss": 0.1059, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.16272137471506226, |
|
"grad_norm": 2.5649735927581787, |
|
"learning_rate": 3.929824561403509e-05, |
|
"loss": 0.0108, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.16342275995090302, |
|
"grad_norm": 389.410400390625, |
|
"learning_rate": 3.9473684210526316e-05, |
|
"loss": 4.1361, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.16412414518674381, |
|
"grad_norm": 97.98786926269531, |
|
"learning_rate": 3.9649122807017545e-05, |
|
"loss": 1.0207, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.1648255304225846, |
|
"grad_norm": 446.8147277832031, |
|
"learning_rate": 3.9824561403508774e-05, |
|
"loss": 2.3678, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.1655269156584254, |
|
"grad_norm": 88.88761901855469, |
|
"learning_rate": 4e-05, |
|
"loss": 0.814, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.16622830089426618, |
|
"grad_norm": 217.3586883544922, |
|
"learning_rate": 4.017543859649123e-05, |
|
"loss": 9.7291, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.16692968613010697, |
|
"grad_norm": 106.2865219116211, |
|
"learning_rate": 4.0350877192982455e-05, |
|
"loss": 1.1716, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.16763107136594774, |
|
"grad_norm": 154.2283477783203, |
|
"learning_rate": 4.0526315789473684e-05, |
|
"loss": 0.6351, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.16833245660178853, |
|
"grad_norm": 37.964447021484375, |
|
"learning_rate": 4.070175438596491e-05, |
|
"loss": 0.3484, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.16903384183762932, |
|
"grad_norm": 24.696714401245117, |
|
"learning_rate": 4.087719298245614e-05, |
|
"loss": 0.2762, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.1697352270734701, |
|
"grad_norm": 82.12918853759766, |
|
"learning_rate": 4.105263157894737e-05, |
|
"loss": 0.9978, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.1704366123093109, |
|
"grad_norm": 41.84835433959961, |
|
"learning_rate": 4.12280701754386e-05, |
|
"loss": 0.4799, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.1711379975451517, |
|
"grad_norm": 45.34370803833008, |
|
"learning_rate": 4.140350877192983e-05, |
|
"loss": 0.4255, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.17183938278099245, |
|
"grad_norm": 41.85076904296875, |
|
"learning_rate": 4.157894736842106e-05, |
|
"loss": 0.616, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.17254076801683324, |
|
"grad_norm": 31.718475341796875, |
|
"learning_rate": 4.1754385964912287e-05, |
|
"loss": 0.2034, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.17324215325267403, |
|
"grad_norm": 149.61940002441406, |
|
"learning_rate": 4.1929824561403516e-05, |
|
"loss": 1.5976, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.17394353848851482, |
|
"grad_norm": 127.43466186523438, |
|
"learning_rate": 4.210526315789474e-05, |
|
"loss": 1.4465, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.1746449237243556, |
|
"grad_norm": 36.885379791259766, |
|
"learning_rate": 4.228070175438597e-05, |
|
"loss": 0.5706, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.1753463089601964, |
|
"grad_norm": 343.46649169921875, |
|
"learning_rate": 4.2456140350877196e-05, |
|
"loss": 5.0778, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.17604769419603716, |
|
"grad_norm": 23.604751586914062, |
|
"learning_rate": 4.2631578947368425e-05, |
|
"loss": 0.1635, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.17674907943187795, |
|
"grad_norm": 37.630043029785156, |
|
"learning_rate": 4.2807017543859654e-05, |
|
"loss": 0.5644, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.17745046466771874, |
|
"grad_norm": 41.07538986206055, |
|
"learning_rate": 4.298245614035088e-05, |
|
"loss": 0.5337, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.17815184990355953, |
|
"grad_norm": 66.12769317626953, |
|
"learning_rate": 4.3157894736842105e-05, |
|
"loss": 0.8491, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.17885323513940032, |
|
"grad_norm": 38.53345489501953, |
|
"learning_rate": 4.3333333333333334e-05, |
|
"loss": 0.4, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.17955462037524111, |
|
"grad_norm": 37.57467269897461, |
|
"learning_rate": 4.350877192982456e-05, |
|
"loss": 0.6272, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.18025600561108188, |
|
"grad_norm": 27.241796493530273, |
|
"learning_rate": 4.368421052631579e-05, |
|
"loss": 0.2307, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.18095739084692267, |
|
"grad_norm": 24.353778839111328, |
|
"learning_rate": 4.3859649122807014e-05, |
|
"loss": 0.1238, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.18165877608276346, |
|
"grad_norm": 5.489215850830078, |
|
"learning_rate": 4.403508771929824e-05, |
|
"loss": 0.031, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.18236016131860425, |
|
"grad_norm": 16.305925369262695, |
|
"learning_rate": 4.421052631578947e-05, |
|
"loss": 0.0398, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.18306154655444504, |
|
"grad_norm": 6.403785228729248, |
|
"learning_rate": 4.43859649122807e-05, |
|
"loss": 0.0304, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.18376293179028583, |
|
"grad_norm": 4.471805572509766, |
|
"learning_rate": 4.456140350877193e-05, |
|
"loss": 0.0077, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.1844643170261266, |
|
"grad_norm": 0.9698334336280823, |
|
"learning_rate": 4.473684210526316e-05, |
|
"loss": 0.0031, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.18516570226196738, |
|
"grad_norm": 0.014244407415390015, |
|
"learning_rate": 4.491228070175439e-05, |
|
"loss": 0.0, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.18586708749780817, |
|
"grad_norm": 121.71415710449219, |
|
"learning_rate": 4.508771929824562e-05, |
|
"loss": 4.2289, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.18656847273364896, |
|
"grad_norm": 0.02348833531141281, |
|
"learning_rate": 4.5263157894736846e-05, |
|
"loss": 0.0001, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.18726985796948975, |
|
"grad_norm": 0.0016702099237591028, |
|
"learning_rate": 4.5438596491228075e-05, |
|
"loss": 0.0, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.18797124320533054, |
|
"grad_norm": 0.0027486609760671854, |
|
"learning_rate": 4.56140350877193e-05, |
|
"loss": 0.0, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.1886726284411713, |
|
"grad_norm": 0.0012574723223224282, |
|
"learning_rate": 4.5789473684210527e-05, |
|
"loss": 0.0, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.1893740136770121, |
|
"grad_norm": 0.002944325562566519, |
|
"learning_rate": 4.5964912280701756e-05, |
|
"loss": 0.0, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.19007539891285288, |
|
"grad_norm": 0.0020754581782966852, |
|
"learning_rate": 4.6140350877192985e-05, |
|
"loss": 0.0, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.19077678414869367, |
|
"grad_norm": 0.008111892268061638, |
|
"learning_rate": 4.6315789473684214e-05, |
|
"loss": 0.0, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.19147816938453446, |
|
"grad_norm": 118.25870513916016, |
|
"learning_rate": 4.649122807017544e-05, |
|
"loss": 3.5798, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.19217955462037525, |
|
"grad_norm": 118.91422271728516, |
|
"learning_rate": 4.666666666666667e-05, |
|
"loss": 2.0933, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.19288093985621602, |
|
"grad_norm": 0.1281914860010147, |
|
"learning_rate": 4.68421052631579e-05, |
|
"loss": 0.0004, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.1935823250920568, |
|
"grad_norm": 0.9166204929351807, |
|
"learning_rate": 4.701754385964913e-05, |
|
"loss": 0.0034, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.1942837103278976, |
|
"grad_norm": 3.7037837505340576, |
|
"learning_rate": 4.719298245614036e-05, |
|
"loss": 0.0179, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.1949850955637384, |
|
"grad_norm": 0.716585099697113, |
|
"learning_rate": 4.736842105263158e-05, |
|
"loss": 0.0028, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.19568648079957918, |
|
"grad_norm": 119.46430969238281, |
|
"learning_rate": 4.754385964912281e-05, |
|
"loss": 1.7075, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.19638786603541997, |
|
"grad_norm": 0.07794260233640671, |
|
"learning_rate": 4.771929824561404e-05, |
|
"loss": 0.0003, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.19708925127126073, |
|
"grad_norm": 119.65841674804688, |
|
"learning_rate": 4.789473684210526e-05, |
|
"loss": 3.3725, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.19779063650710152, |
|
"grad_norm": 0.06007947400212288, |
|
"learning_rate": 4.807017543859649e-05, |
|
"loss": 0.0002, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.1984920217429423, |
|
"grad_norm": 0.33029523491859436, |
|
"learning_rate": 4.824561403508772e-05, |
|
"loss": 0.0013, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.1991934069787831, |
|
"grad_norm": 0.19337859749794006, |
|
"learning_rate": 4.842105263157895e-05, |
|
"loss": 0.0007, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.1998947922146239, |
|
"grad_norm": 0.4252503514289856, |
|
"learning_rate": 4.859649122807018e-05, |
|
"loss": 0.0017, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.20059617745046468, |
|
"grad_norm": 0.07317744940519333, |
|
"learning_rate": 4.8771929824561406e-05, |
|
"loss": 0.0003, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.20129756268630544, |
|
"grad_norm": 0.01806594245135784, |
|
"learning_rate": 4.8947368421052635e-05, |
|
"loss": 0.0001, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.20199894792214623, |
|
"grad_norm": 106.39412689208984, |
|
"learning_rate": 4.912280701754386e-05, |
|
"loss": 3.9844, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.20270033315798702, |
|
"grad_norm": 0.008989217691123486, |
|
"learning_rate": 4.9298245614035086e-05, |
|
"loss": 0.0, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.2034017183938278, |
|
"grad_norm": 0.05597059056162834, |
|
"learning_rate": 4.9473684210526315e-05, |
|
"loss": 0.0002, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.2041031036296686, |
|
"grad_norm": 113.05389404296875, |
|
"learning_rate": 4.9649122807017544e-05, |
|
"loss": 2.5036, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.2048044888655094, |
|
"grad_norm": 0.39913010597229004, |
|
"learning_rate": 4.9824561403508773e-05, |
|
"loss": 0.0015, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.20550587410135016, |
|
"grad_norm": 0.944106936454773, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0038, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.20620725933719095, |
|
"grad_norm": 108.76079559326172, |
|
"learning_rate": 4.9999995795715716e-05, |
|
"loss": 1.4773, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.20690864457303174, |
|
"grad_norm": 1.2869099378585815, |
|
"learning_rate": 4.999998318286425e-05, |
|
"loss": 0.0054, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.20761002980887253, |
|
"grad_norm": 0.7701173424720764, |
|
"learning_rate": 4.999996216144987e-05, |
|
"loss": 0.0029, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.20831141504471332, |
|
"grad_norm": 104.06690216064453, |
|
"learning_rate": 4.9999932731479625e-05, |
|
"loss": 1.2931, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.2090128002805541, |
|
"grad_norm": 234.58531188964844, |
|
"learning_rate": 4.999989489296344e-05, |
|
"loss": 4.5688, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.20971418551639487, |
|
"grad_norm": 108.04949188232422, |
|
"learning_rate": 4.999984864591401e-05, |
|
"loss": 1.5544, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.21041557075223566, |
|
"grad_norm": 1.8640260696411133, |
|
"learning_rate": 4.999979399034691e-05, |
|
"loss": 0.0078, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.21111695598807645, |
|
"grad_norm": 3.089526414871216, |
|
"learning_rate": 4.999973092628052e-05, |
|
"loss": 0.0138, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.21181834122391724, |
|
"grad_norm": 1.4183728694915771, |
|
"learning_rate": 4.999965945373605e-05, |
|
"loss": 0.006, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.21251972645975803, |
|
"grad_norm": 0.9733534455299377, |
|
"learning_rate": 4.9999579572737533e-05, |
|
"loss": 0.0038, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.21322111169559882, |
|
"grad_norm": 121.11236572265625, |
|
"learning_rate": 4.9999491283311836e-05, |
|
"loss": 2.2543, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.21392249693143958, |
|
"grad_norm": 117.718017578125, |
|
"learning_rate": 4.999939458548868e-05, |
|
"loss": 2.5007, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.21462388216728037, |
|
"grad_norm": 0.013796065002679825, |
|
"learning_rate": 4.9999289479300557e-05, |
|
"loss": 0.0, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.21532526740312116, |
|
"grad_norm": 116.46581268310547, |
|
"learning_rate": 4.999917596478283e-05, |
|
"loss": 2.6005, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.21602665263896195, |
|
"grad_norm": 0.3713426887989044, |
|
"learning_rate": 4.999905404197368e-05, |
|
"loss": 0.0014, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.21672803787480274, |
|
"grad_norm": 0.44403278827667236, |
|
"learning_rate": 4.999892371091411e-05, |
|
"loss": 0.0017, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.21742942311064353, |
|
"grad_norm": 139.35816955566406, |
|
"learning_rate": 4.999878497164797e-05, |
|
"loss": 11.1347, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.2181308083464843, |
|
"grad_norm": 0.20512554049491882, |
|
"learning_rate": 4.999863782422191e-05, |
|
"loss": 0.0007, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.21883219358232509, |
|
"grad_norm": 88.1954345703125, |
|
"learning_rate": 4.9998482268685434e-05, |
|
"loss": 0.843, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.21953357881816588, |
|
"grad_norm": 93.83118438720703, |
|
"learning_rate": 4.999831830509084e-05, |
|
"loss": 1.0804, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.22023496405400667, |
|
"grad_norm": 5.042333602905273, |
|
"learning_rate": 4.99981459334933e-05, |
|
"loss": 0.0256, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.22093634928984746, |
|
"grad_norm": 7.477559566497803, |
|
"learning_rate": 4.9997965153950775e-05, |
|
"loss": 0.0402, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.22163773452568825, |
|
"grad_norm": 156.78961181640625, |
|
"learning_rate": 4.999777596652408e-05, |
|
"loss": 6.0824, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.222339119761529, |
|
"grad_norm": 9.518619537353516, |
|
"learning_rate": 4.999757837127683e-05, |
|
"loss": 0.0577, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.2230405049973698, |
|
"grad_norm": 2.860729455947876, |
|
"learning_rate": 4.9997372368275506e-05, |
|
"loss": 0.0135, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.2237418902332106, |
|
"grad_norm": 1.1225600242614746, |
|
"learning_rate": 4.999715795758938e-05, |
|
"loss": 0.0049, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.22444327546905138, |
|
"grad_norm": 0.12374955415725708, |
|
"learning_rate": 4.9996935139290575e-05, |
|
"loss": 0.0004, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.22514466070489217, |
|
"grad_norm": 112.18521881103516, |
|
"learning_rate": 4.999670391345403e-05, |
|
"loss": 3.4372, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.22584604594073296, |
|
"grad_norm": 0.0004930093418806791, |
|
"learning_rate": 4.999646428015752e-05, |
|
"loss": 0.0, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.22654743117657372, |
|
"grad_norm": 0.00027333476464264095, |
|
"learning_rate": 4.9996216239481643e-05, |
|
"loss": 0.0, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.2272488164124145, |
|
"grad_norm": 0.0001735862751957029, |
|
"learning_rate": 4.999595979150982e-05, |
|
"loss": 0.0, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.2279502016482553, |
|
"grad_norm": 2.878743180190213e-05, |
|
"learning_rate": 4.9995694936328316e-05, |
|
"loss": 0.0, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.2286515868840961, |
|
"grad_norm": 109.58158111572266, |
|
"learning_rate": 4.99954216740262e-05, |
|
"loss": 4.991, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.22935297211993688, |
|
"grad_norm": 113.2159194946289, |
|
"learning_rate": 4.9995140004695396e-05, |
|
"loss": 4.7235, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.23005435735577767, |
|
"grad_norm": 0.00015405479643959552, |
|
"learning_rate": 4.999484992843063e-05, |
|
"loss": 0.0, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.23075574259161843, |
|
"grad_norm": 0.011504331603646278, |
|
"learning_rate": 4.999455144532947e-05, |
|
"loss": 0.0, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.23145712782745922, |
|
"grad_norm": 0.001587590086273849, |
|
"learning_rate": 4.9994244555492315e-05, |
|
"loss": 0.0, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.23215851306330001, |
|
"grad_norm": 0.0038678112905472517, |
|
"learning_rate": 4.999392925902238e-05, |
|
"loss": 0.0, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.2328598982991408, |
|
"grad_norm": 0.044445931911468506, |
|
"learning_rate": 4.999360555602571e-05, |
|
"loss": 0.0001, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.2335612835349816, |
|
"grad_norm": 108.20228576660156, |
|
"learning_rate": 4.999327344661118e-05, |
|
"loss": 2.0004, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.23426266877082239, |
|
"grad_norm": 0.17047792673110962, |
|
"learning_rate": 4.9992932930890505e-05, |
|
"loss": 0.0005, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.23496405400666315, |
|
"grad_norm": 0.39230290055274963, |
|
"learning_rate": 4.999258400897819e-05, |
|
"loss": 0.0016, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.23566543924250394, |
|
"grad_norm": 191.8765106201172, |
|
"learning_rate": 4.999222668099163e-05, |
|
"loss": 2.4042, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.23636682447834473, |
|
"grad_norm": 87.04983520507812, |
|
"learning_rate": 4.999186094705097e-05, |
|
"loss": 0.9342, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.23706820971418552, |
|
"grad_norm": 14.748551368713379, |
|
"learning_rate": 4.999148680727925e-05, |
|
"loss": 0.1165, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.2377695949500263, |
|
"grad_norm": 16.103792190551758, |
|
"learning_rate": 4.999110426180229e-05, |
|
"loss": 0.11, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.2384709801858671, |
|
"grad_norm": 4.507053375244141, |
|
"learning_rate": 4.999071331074877e-05, |
|
"loss": 0.1421, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.23917236542170786, |
|
"grad_norm": 8.822351455688477, |
|
"learning_rate": 4.999031395425018e-05, |
|
"loss": 0.2372, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.23987375065754865, |
|
"grad_norm": 19.92313575744629, |
|
"learning_rate": 4.998990619244084e-05, |
|
"loss": 0.1537, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.24057513589338944, |
|
"grad_norm": 12.00966739654541, |
|
"learning_rate": 4.998949002545789e-05, |
|
"loss": 0.0872, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.24127652112923023, |
|
"grad_norm": 4.180269718170166, |
|
"learning_rate": 4.998906545344132e-05, |
|
"loss": 0.0218, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.24197790636507102, |
|
"grad_norm": 0.596821129322052, |
|
"learning_rate": 4.998863247653392e-05, |
|
"loss": 0.0025, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.2426792916009118, |
|
"grad_norm": 0.4522903263568878, |
|
"learning_rate": 4.998819109488132e-05, |
|
"loss": 0.0013, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.24338067683675257, |
|
"grad_norm": 0.0007598726078867912, |
|
"learning_rate": 4.998774130863199e-05, |
|
"loss": 0.0, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.24408206207259336, |
|
"grad_norm": 8.103512664092705e-05, |
|
"learning_rate": 4.998728311793719e-05, |
|
"loss": 0.0, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.24478344730843415, |
|
"grad_norm": 105.80138397216797, |
|
"learning_rate": 4.998681652295104e-05, |
|
"loss": 4.8223, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.24548483254427494, |
|
"grad_norm": 127.92369079589844, |
|
"learning_rate": 4.998634152383047e-05, |
|
"loss": 7.0698, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.24618621778011573, |
|
"grad_norm": 2.636051704030251e-06, |
|
"learning_rate": 4.998585812073525e-05, |
|
"loss": 0.0, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.24688760301595652, |
|
"grad_norm": 2.710710077735712e-06, |
|
"learning_rate": 4.9985366313827975e-05, |
|
"loss": 0.0, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.2475889882517973, |
|
"grad_norm": 103.96717834472656, |
|
"learning_rate": 4.998486610327405e-05, |
|
"loss": 6.485, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.24829037348763808, |
|
"grad_norm": 103.35551452636719, |
|
"learning_rate": 4.9984357489241715e-05, |
|
"loss": 5.7037, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.24899175872347887, |
|
"grad_norm": 100.56452178955078, |
|
"learning_rate": 4.998384047190204e-05, |
|
"loss": 4.1908, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.24969314395931966, |
|
"grad_norm": 200.88751220703125, |
|
"learning_rate": 4.998331505142893e-05, |
|
"loss": 4.8984, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.2503945291951604, |
|
"grad_norm": 77.69247436523438, |
|
"learning_rate": 4.99827812279991e-05, |
|
"loss": 0.9449, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.2510959144310012, |
|
"grad_norm": 26.622453689575195, |
|
"learning_rate": 4.9982239001792095e-05, |
|
"loss": 0.3613, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.251797299666842, |
|
"grad_norm": 26.442935943603516, |
|
"learning_rate": 4.9981688372990284e-05, |
|
"loss": 0.5424, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.2524986849026828, |
|
"grad_norm": 39.277618408203125, |
|
"learning_rate": 4.9981129341778886e-05, |
|
"loss": 0.9456, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.2532000701385236, |
|
"grad_norm": 43.96021270751953, |
|
"learning_rate": 4.9980561908345916e-05, |
|
"loss": 0.6765, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.25390145537436437, |
|
"grad_norm": 14.254291534423828, |
|
"learning_rate": 4.997998607288222e-05, |
|
"loss": 0.1276, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.25460284061020516, |
|
"grad_norm": 0.6544491052627563, |
|
"learning_rate": 4.9979401835581476e-05, |
|
"loss": 0.0032, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.25530422584604595, |
|
"grad_norm": 0.011080354452133179, |
|
"learning_rate": 4.99788091966402e-05, |
|
"loss": 0.0, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.25600561108188674, |
|
"grad_norm": 284.8055114746094, |
|
"learning_rate": 4.997820815625771e-05, |
|
"loss": 7.6058, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.25670699631772753, |
|
"grad_norm": 89.02140045166016, |
|
"learning_rate": 4.997759871463618e-05, |
|
"loss": 5.079, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.2574083815535683, |
|
"grad_norm": 4.7222201828844845e-05, |
|
"learning_rate": 4.997698087198056e-05, |
|
"loss": 0.0, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.25810976678940906, |
|
"grad_norm": 85.66144561767578, |
|
"learning_rate": 4.997635462849869e-05, |
|
"loss": 4.8183, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.25881115202524985, |
|
"grad_norm": 0.0015366720035672188, |
|
"learning_rate": 4.997571998440118e-05, |
|
"loss": 0.0, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.25951253726109064, |
|
"grad_norm": 100.14698028564453, |
|
"learning_rate": 4.99750769399015e-05, |
|
"loss": 2.2663, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.2602139224969314, |
|
"grad_norm": 0.4829590618610382, |
|
"learning_rate": 4.997442549521592e-05, |
|
"loss": 0.0021, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.2609153077327722, |
|
"grad_norm": 70.44209289550781, |
|
"learning_rate": 4.9973765650563564e-05, |
|
"loss": 0.7532, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.261616692968613, |
|
"grad_norm": 272.1163024902344, |
|
"learning_rate": 4.9973097406166355e-05, |
|
"loss": 3.9935, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.2623180782044538, |
|
"grad_norm": 25.12706184387207, |
|
"learning_rate": 4.997242076224906e-05, |
|
"loss": 0.3738, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.2630194634402946, |
|
"grad_norm": 35.15785598754883, |
|
"learning_rate": 4.997173571903926e-05, |
|
"loss": 0.6808, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.2637208486761354, |
|
"grad_norm": 27.46406364440918, |
|
"learning_rate": 4.9971042276767366e-05, |
|
"loss": 0.2568, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.26442223391197617, |
|
"grad_norm": 6.682714462280273, |
|
"learning_rate": 4.997034043566661e-05, |
|
"loss": 0.0374, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.26512361914781696, |
|
"grad_norm": 0.057287223637104034, |
|
"learning_rate": 4.9969630195973045e-05, |
|
"loss": 0.0001, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.26582500438365775, |
|
"grad_norm": 0.006871811114251614, |
|
"learning_rate": 4.9968911557925564e-05, |
|
"loss": 0.0, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.2665263896194985, |
|
"grad_norm": 0.0037054684944450855, |
|
"learning_rate": 4.996818452176587e-05, |
|
"loss": 0.0, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.2672277748553393, |
|
"grad_norm": 123.15022277832031, |
|
"learning_rate": 4.9967449087738505e-05, |
|
"loss": 5.8346, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.26792916009118006, |
|
"grad_norm": 121.81306457519531, |
|
"learning_rate": 4.996670525609082e-05, |
|
"loss": 4.3778, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.26863054532702085, |
|
"grad_norm": 0.00026572938077151775, |
|
"learning_rate": 4.9965953027072996e-05, |
|
"loss": 0.0, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.26933193056286164, |
|
"grad_norm": 198.09776306152344, |
|
"learning_rate": 4.996519240093806e-05, |
|
"loss": 5.2574, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.27003331579870243, |
|
"grad_norm": 130.21214294433594, |
|
"learning_rate": 4.996442337794181e-05, |
|
"loss": 3.217, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.2707347010345432, |
|
"grad_norm": 133.94178771972656, |
|
"learning_rate": 4.996364595834291e-05, |
|
"loss": 1.3412, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.271436086270384, |
|
"grad_norm": 12.223856925964355, |
|
"learning_rate": 4.9962860142402853e-05, |
|
"loss": 0.0924, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.2721374715062248, |
|
"grad_norm": 21.94683074951172, |
|
"learning_rate": 4.9962065930385934e-05, |
|
"loss": 0.1824, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.2728388567420656, |
|
"grad_norm": 46.802406311035156, |
|
"learning_rate": 4.996126332255929e-05, |
|
"loss": 0.3585, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.2735402419779064, |
|
"grad_norm": 7.21342134475708, |
|
"learning_rate": 4.9960452319192866e-05, |
|
"loss": 0.1045, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.2742416272137472, |
|
"grad_norm": 33.5169677734375, |
|
"learning_rate": 4.995963292055943e-05, |
|
"loss": 0.248, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.2749430124495879, |
|
"grad_norm": 16.911136627197266, |
|
"learning_rate": 4.995880512693458e-05, |
|
"loss": 0.0876, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.2756443976854287, |
|
"grad_norm": 206.7276153564453, |
|
"learning_rate": 4.995796893859675e-05, |
|
"loss": 1.6484, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.2763457829212695, |
|
"grad_norm": 94.29922485351562, |
|
"learning_rate": 4.9957124355827186e-05, |
|
"loss": 0.8745, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.2770471681571103, |
|
"grad_norm": 156.6600799560547, |
|
"learning_rate": 4.9956271378909945e-05, |
|
"loss": 0.9049, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.27774855339295107, |
|
"grad_norm": 6.1268744468688965, |
|
"learning_rate": 4.995541000813193e-05, |
|
"loss": 0.0343, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.27844993862879186, |
|
"grad_norm": 1.731136679649353, |
|
"learning_rate": 4.995454024378285e-05, |
|
"loss": 0.0075, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.27915132386463265, |
|
"grad_norm": 144.5174560546875, |
|
"learning_rate": 4.995366208615526e-05, |
|
"loss": 2.6072, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.27985270910047344, |
|
"grad_norm": 0.19711287319660187, |
|
"learning_rate": 4.9952775535544494e-05, |
|
"loss": 0.0007, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.28055409433631423, |
|
"grad_norm": 0.13065628707408905, |
|
"learning_rate": 4.9951880592248754e-05, |
|
"loss": 0.0004, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.281255479572155, |
|
"grad_norm": 125.28437042236328, |
|
"learning_rate": 4.995097725656905e-05, |
|
"loss": 1.8792, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.2819568648079958, |
|
"grad_norm": 111.38099670410156, |
|
"learning_rate": 4.9950065528809214e-05, |
|
"loss": 2.548, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.2826582500438366, |
|
"grad_norm": 1.2473199367523193, |
|
"learning_rate": 4.994914540927589e-05, |
|
"loss": 0.0042, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.28335963527967734, |
|
"grad_norm": 0.7474699020385742, |
|
"learning_rate": 4.994821689827856e-05, |
|
"loss": 0.0026, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.2840610205155181, |
|
"grad_norm": 0.14196322858333588, |
|
"learning_rate": 4.99472799961295e-05, |
|
"loss": 0.0005, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.2847624057513589, |
|
"grad_norm": 0.6420115828514099, |
|
"learning_rate": 4.994633470314387e-05, |
|
"loss": 0.0026, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.2854637909871997, |
|
"grad_norm": 213.8779296875, |
|
"learning_rate": 4.994538101963958e-05, |
|
"loss": 2.1989, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.2861651762230405, |
|
"grad_norm": 108.08120727539062, |
|
"learning_rate": 4.994441894593741e-05, |
|
"loss": 3.9965, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.2868665614588813, |
|
"grad_norm": 0.40015965700149536, |
|
"learning_rate": 4.994344848236094e-05, |
|
"loss": 0.0016, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.2875679466947221, |
|
"grad_norm": 125.80431365966797, |
|
"learning_rate": 4.9942469629236585e-05, |
|
"loss": 2.3139, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.28826933193056287, |
|
"grad_norm": 101.27831268310547, |
|
"learning_rate": 4.994148238689357e-05, |
|
"loss": 2.1065, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.28897071716640366, |
|
"grad_norm": 3.526338577270508, |
|
"learning_rate": 4.9940486755663946e-05, |
|
"loss": 0.015, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.28967210240224445, |
|
"grad_norm": 2.501554489135742, |
|
"learning_rate": 4.993948273588259e-05, |
|
"loss": 0.0123, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.29037348763808524, |
|
"grad_norm": 75.39434814453125, |
|
"learning_rate": 4.993847032788719e-05, |
|
"loss": 0.7107, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.291074872873926, |
|
"grad_norm": 3.751978635787964, |
|
"learning_rate": 4.993744953201827e-05, |
|
"loss": 0.0183, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.29177625810976676, |
|
"grad_norm": 5.31666374206543, |
|
"learning_rate": 4.9936420348619164e-05, |
|
"loss": 0.0385, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.29247764334560755, |
|
"grad_norm": 104.77239227294922, |
|
"learning_rate": 4.9935382778036036e-05, |
|
"loss": 1.7353, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.29317902858144834, |
|
"grad_norm": 92.44829559326172, |
|
"learning_rate": 4.993433682061786e-05, |
|
"loss": 1.1895, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.29388041381728913, |
|
"grad_norm": 89.00714874267578, |
|
"learning_rate": 4.993328247671643e-05, |
|
"loss": 1.2323, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.2945817990531299, |
|
"grad_norm": 172.867431640625, |
|
"learning_rate": 4.993221974668637e-05, |
|
"loss": 1.7303, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.2952831842889707, |
|
"grad_norm": 19.69775390625, |
|
"learning_rate": 4.9931148630885127e-05, |
|
"loss": 0.1678, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.2959845695248115, |
|
"grad_norm": 22.828763961791992, |
|
"learning_rate": 4.9930069129672965e-05, |
|
"loss": 0.3184, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.2966859547606523, |
|
"grad_norm": 19.460153579711914, |
|
"learning_rate": 4.992898124341296e-05, |
|
"loss": 0.2748, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.2973873399964931, |
|
"grad_norm": 2.9695911407470703, |
|
"learning_rate": 4.9927884972471015e-05, |
|
"loss": 0.1925, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.2980887252323339, |
|
"grad_norm": 12.33647346496582, |
|
"learning_rate": 4.9926780317215854e-05, |
|
"loss": 0.1908, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.29879011046817466, |
|
"grad_norm": 2.8421826362609863, |
|
"learning_rate": 4.992566727801902e-05, |
|
"loss": 0.101, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.29949149570401545, |
|
"grad_norm": 171.8658447265625, |
|
"learning_rate": 4.9924545855254875e-05, |
|
"loss": 3.9311, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.30019288093985624, |
|
"grad_norm": 83.66468048095703, |
|
"learning_rate": 4.992341604930061e-05, |
|
"loss": 1.6806, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.300894266175697, |
|
"grad_norm": 0.7609736323356628, |
|
"learning_rate": 4.9922277860536216e-05, |
|
"loss": 0.0041, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.30159565141153777, |
|
"grad_norm": 4.077641487121582, |
|
"learning_rate": 4.9921131289344514e-05, |
|
"loss": 0.0418, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.30229703664737856, |
|
"grad_norm": 3.2341866493225098, |
|
"learning_rate": 4.9919976336111155e-05, |
|
"loss": 0.0293, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.30299842188321935, |
|
"grad_norm": 148.74095153808594, |
|
"learning_rate": 4.991881300122458e-05, |
|
"loss": 3.3951, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.30369980711906014, |
|
"grad_norm": 77.130126953125, |
|
"learning_rate": 4.99176412850761e-05, |
|
"loss": 2.9407, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.30440119235490093, |
|
"grad_norm": 162.98777770996094, |
|
"learning_rate": 4.991646118805978e-05, |
|
"loss": 4.3504, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.3051025775907417, |
|
"grad_norm": 84.4205551147461, |
|
"learning_rate": 4.991527271057256e-05, |
|
"loss": 1.7142, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.3058039628265825, |
|
"grad_norm": 0.8636782765388489, |
|
"learning_rate": 4.991407585301416e-05, |
|
"loss": 0.0053, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.3065053480624233, |
|
"grad_norm": 1.0418014526367188, |
|
"learning_rate": 4.9912870615787144e-05, |
|
"loss": 0.0059, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.3072067332982641, |
|
"grad_norm": 0.7050487399101257, |
|
"learning_rate": 4.991165699929688e-05, |
|
"loss": 0.0041, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.3079081185341049, |
|
"grad_norm": 0.3087022304534912, |
|
"learning_rate": 4.991043500395157e-05, |
|
"loss": 0.0017, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.30860950376994567, |
|
"grad_norm": 0.09886661171913147, |
|
"learning_rate": 4.990920463016221e-05, |
|
"loss": 0.0005, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.3093108890057864, |
|
"grad_norm": 0.03563763573765755, |
|
"learning_rate": 4.9907965878342634e-05, |
|
"loss": 0.0002, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.3100122742416272, |
|
"grad_norm": 79.55757141113281, |
|
"learning_rate": 4.990671874890948e-05, |
|
"loss": 2.966, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.310713659477468, |
|
"grad_norm": 0.6562897562980652, |
|
"learning_rate": 4.990546324228222e-05, |
|
"loss": 0.0022, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.3114150447133088, |
|
"grad_norm": 74.84564208984375, |
|
"learning_rate": 4.990419935888313e-05, |
|
"loss": 3.0154, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.31211642994914957, |
|
"grad_norm": 0.03038967400789261, |
|
"learning_rate": 4.9902927099137305e-05, |
|
"loss": 0.0002, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.31281781518499036, |
|
"grad_norm": 0.023526914417743683, |
|
"learning_rate": 4.9901646463472674e-05, |
|
"loss": 0.0001, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.31351920042083115, |
|
"grad_norm": 0.08529358357191086, |
|
"learning_rate": 4.990035745231995e-05, |
|
"loss": 0.0005, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.31422058565667194, |
|
"grad_norm": 0.11940905451774597, |
|
"learning_rate": 4.989906006611269e-05, |
|
"loss": 0.0007, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.3149219708925127, |
|
"grad_norm": 99.68978118896484, |
|
"learning_rate": 4.989775430528727e-05, |
|
"loss": 1.6958, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.3156233561283535, |
|
"grad_norm": 0.3344658315181732, |
|
"learning_rate": 4.9896440170282864e-05, |
|
"loss": 0.002, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.3163247413641943, |
|
"grad_norm": 0.7184603810310364, |
|
"learning_rate": 4.9895117661541476e-05, |
|
"loss": 0.0043, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.3170261266000351, |
|
"grad_norm": 0.5113491415977478, |
|
"learning_rate": 4.989378677950791e-05, |
|
"loss": 0.0031, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.31772751183587583, |
|
"grad_norm": 0.49223193526268005, |
|
"learning_rate": 4.989244752462982e-05, |
|
"loss": 0.0029, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.3184288970717166, |
|
"grad_norm": 64.00535583496094, |
|
"learning_rate": 4.989109989735763e-05, |
|
"loss": 1.2629, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.3191302823075574, |
|
"grad_norm": 0.18930155038833618, |
|
"learning_rate": 4.988974389814463e-05, |
|
"loss": 0.0011, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.3198316675433982, |
|
"grad_norm": 0.3088482916355133, |
|
"learning_rate": 4.9888379527446885e-05, |
|
"loss": 0.0018, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.320533052779239, |
|
"grad_norm": 71.42636108398438, |
|
"learning_rate": 4.988700678572329e-05, |
|
"loss": 1.9691, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.3212344380150798, |
|
"grad_norm": 0.20487411320209503, |
|
"learning_rate": 4.988562567343557e-05, |
|
"loss": 0.0012, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.32193582325092057, |
|
"grad_norm": 73.37615203857422, |
|
"learning_rate": 4.988423619104823e-05, |
|
"loss": 1.8636, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.32263720848676136, |
|
"grad_norm": 66.4114761352539, |
|
"learning_rate": 4.988283833902863e-05, |
|
"loss": 1.3073, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.32333859372260215, |
|
"grad_norm": 4.2518744468688965, |
|
"learning_rate": 4.988143211784693e-05, |
|
"loss": 0.0318, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.32403997895844294, |
|
"grad_norm": 4.782449245452881, |
|
"learning_rate": 4.9880017527976083e-05, |
|
"loss": 0.0384, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.32474136419428373, |
|
"grad_norm": 22.517131805419922, |
|
"learning_rate": 4.9878594569891896e-05, |
|
"loss": 0.2973, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.3254427494301245, |
|
"grad_norm": 5.63627290725708, |
|
"learning_rate": 4.987716324407296e-05, |
|
"loss": 0.0496, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.32614413466596526, |
|
"grad_norm": 6.364257335662842, |
|
"learning_rate": 4.9875723551000696e-05, |
|
"loss": 0.0578, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.32684551990180605, |
|
"grad_norm": 147.82330322265625, |
|
"learning_rate": 4.987427549115933e-05, |
|
"loss": 1.8188, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.32754690513764684, |
|
"grad_norm": 3.7355575561523438, |
|
"learning_rate": 4.987281906503591e-05, |
|
"loss": 0.0274, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.32824829037348763, |
|
"grad_norm": 0.6107962131500244, |
|
"learning_rate": 4.987135427312029e-05, |
|
"loss": 0.0038, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.3289496756093284, |
|
"grad_norm": 61.26199722290039, |
|
"learning_rate": 4.9869881115905146e-05, |
|
"loss": 0.8915, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.3296510608451692, |
|
"grad_norm": 0.3030995726585388, |
|
"learning_rate": 4.986839959388596e-05, |
|
"loss": 0.0018, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.33035244608101, |
|
"grad_norm": 0.12271957844495773, |
|
"learning_rate": 4.986690970756104e-05, |
|
"loss": 0.0007, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.3310538313168508, |
|
"grad_norm": 71.7486801147461, |
|
"learning_rate": 4.9865411457431485e-05, |
|
"loss": 1.8673, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.3317552165526916, |
|
"grad_norm": 69.17851257324219, |
|
"learning_rate": 4.986390484400124e-05, |
|
"loss": 1.6587, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.33245660178853237, |
|
"grad_norm": 0.18208004534244537, |
|
"learning_rate": 4.9862389867777023e-05, |
|
"loss": 0.001, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.33315798702437316, |
|
"grad_norm": 122.65370178222656, |
|
"learning_rate": 4.986086652926839e-05, |
|
"loss": 0.9068, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.33385937226021395, |
|
"grad_norm": 0.6119470596313477, |
|
"learning_rate": 4.9859334828987715e-05, |
|
"loss": 0.0039, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.3345607574960547, |
|
"grad_norm": 4.483019828796387, |
|
"learning_rate": 4.985779476745017e-05, |
|
"loss": 0.0343, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.3352621427318955, |
|
"grad_norm": 0.7592105269432068, |
|
"learning_rate": 4.985624634517374e-05, |
|
"loss": 0.005, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.33596352796773626, |
|
"grad_norm": 0.6566614508628845, |
|
"learning_rate": 4.985468956267922e-05, |
|
"loss": 0.0037, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.33666491320357705, |
|
"grad_norm": 67.39823150634766, |
|
"learning_rate": 4.985312442049023e-05, |
|
"loss": 1.7049, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.33736629843941784, |
|
"grad_norm": 0.13803933560848236, |
|
"learning_rate": 4.98515509191332e-05, |
|
"loss": 0.0008, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.33806768367525863, |
|
"grad_norm": 72.15523529052734, |
|
"learning_rate": 4.984996905913736e-05, |
|
"loss": 1.6634, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.3387690689110994, |
|
"grad_norm": 0.10033612698316574, |
|
"learning_rate": 4.984837884103475e-05, |
|
"loss": 0.0005, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.3394704541469402, |
|
"grad_norm": 62.519317626953125, |
|
"learning_rate": 4.984678026536024e-05, |
|
"loss": 1.2577, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.340171839382781, |
|
"grad_norm": 0.36229056119918823, |
|
"learning_rate": 4.984517333265148e-05, |
|
"loss": 0.0023, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.3408732246186218, |
|
"grad_norm": 0.8776547312736511, |
|
"learning_rate": 4.9843558043448985e-05, |
|
"loss": 0.0056, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.3415746098544626, |
|
"grad_norm": 108.73171997070312, |
|
"learning_rate": 4.984193439829602e-05, |
|
"loss": 2.7267, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.3422759950903034, |
|
"grad_norm": 0.6344736814498901, |
|
"learning_rate": 4.984030239773868e-05, |
|
"loss": 0.0041, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.3429773803261441, |
|
"grad_norm": 113.62140655517578, |
|
"learning_rate": 4.98386620423259e-05, |
|
"loss": 1.7271, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.3436787655619849, |
|
"grad_norm": 1.618384599685669, |
|
"learning_rate": 4.983701333260938e-05, |
|
"loss": 0.011, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.3443801507978257, |
|
"grad_norm": 138.78854370117188, |
|
"learning_rate": 4.9835356269143676e-05, |
|
"loss": 1.9143, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.3450815360336665, |
|
"grad_norm": 49.62142562866211, |
|
"learning_rate": 4.98336908524861e-05, |
|
"loss": 0.753, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.34578292126950727, |
|
"grad_norm": 3.6228299140930176, |
|
"learning_rate": 4.9832017083196816e-05, |
|
"loss": 0.0277, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.34648430650534806, |
|
"grad_norm": 24.915790557861328, |
|
"learning_rate": 4.98303349618388e-05, |
|
"loss": 0.356, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.34718569174118885, |
|
"grad_norm": 7.880053997039795, |
|
"learning_rate": 4.9828644488977795e-05, |
|
"loss": 0.0727, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.34788707697702964, |
|
"grad_norm": 12.95199203491211, |
|
"learning_rate": 4.982694566518239e-05, |
|
"loss": 0.25, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.34858846221287043, |
|
"grad_norm": 2.059384822845459, |
|
"learning_rate": 4.982523849102397e-05, |
|
"loss": 0.2069, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.3492898474487112, |
|
"grad_norm": 17.132720947265625, |
|
"learning_rate": 4.982352296707675e-05, |
|
"loss": 0.2093, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.349991232684552, |
|
"grad_norm": 12.591998100280762, |
|
"learning_rate": 4.98217990939177e-05, |
|
"loss": 0.1889, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.3506926179203928, |
|
"grad_norm": 12.747396469116211, |
|
"learning_rate": 4.982006687212666e-05, |
|
"loss": 0.1755, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.35139400315623354, |
|
"grad_norm": 77.32592010498047, |
|
"learning_rate": 4.981832630228624e-05, |
|
"loss": 0.9517, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.3520953883920743, |
|
"grad_norm": 8.452153205871582, |
|
"learning_rate": 4.9816577384981866e-05, |
|
"loss": 0.0913, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.3527967736279151, |
|
"grad_norm": 2.1267127990722656, |
|
"learning_rate": 4.9814820120801776e-05, |
|
"loss": 0.0178, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.3534981588637559, |
|
"grad_norm": 0.8161488771438599, |
|
"learning_rate": 4.981305451033701e-05, |
|
"loss": 0.0058, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.3541995440995967, |
|
"grad_norm": 0.10489612072706223, |
|
"learning_rate": 4.981128055418143e-05, |
|
"loss": 0.0006, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.3549009293354375, |
|
"grad_norm": 69.51275634765625, |
|
"learning_rate": 4.9809498252931674e-05, |
|
"loss": 2.62, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.3556023145712783, |
|
"grad_norm": 0.004451240878552198, |
|
"learning_rate": 4.980770760718721e-05, |
|
"loss": 0.0, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.35630369980711907, |
|
"grad_norm": 67.6517105102539, |
|
"learning_rate": 4.9805908617550335e-05, |
|
"loss": 3.25, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.35700508504295986, |
|
"grad_norm": 0.016789492219686508, |
|
"learning_rate": 4.980410128462609e-05, |
|
"loss": 0.0001, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.35770647027880065, |
|
"grad_norm": 0.025487884879112244, |
|
"learning_rate": 4.980228560902239e-05, |
|
"loss": 0.0001, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.35840785551464144, |
|
"grad_norm": 0.042408283799886703, |
|
"learning_rate": 4.98004615913499e-05, |
|
"loss": 0.0002, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.35910924075048223, |
|
"grad_norm": 0.038653161376714706, |
|
"learning_rate": 4.979862923222212e-05, |
|
"loss": 0.0002, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.35981062598632296, |
|
"grad_norm": 0.03510050103068352, |
|
"learning_rate": 4.9796788532255356e-05, |
|
"loss": 0.0002, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.36051201122216375, |
|
"grad_norm": 0.008441315963864326, |
|
"learning_rate": 4.979493949206872e-05, |
|
"loss": 0.0, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.36121339645800454, |
|
"grad_norm": 0.006373463664203882, |
|
"learning_rate": 4.979308211228411e-05, |
|
"loss": 0.0, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.36191478169384533, |
|
"grad_norm": 79.59523010253906, |
|
"learning_rate": 4.979121639352625e-05, |
|
"loss": 2.9487, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.3626161669296861, |
|
"grad_norm": 153.0602264404297, |
|
"learning_rate": 4.9789342336422665e-05, |
|
"loss": 3.6533, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.3633175521655269, |
|
"grad_norm": 0.7085576057434082, |
|
"learning_rate": 4.9787459941603666e-05, |
|
"loss": 0.0042, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.3640189374013677, |
|
"grad_norm": 155.89154052734375, |
|
"learning_rate": 4.9785569209702395e-05, |
|
"loss": 7.9018, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.3647203226372085, |
|
"grad_norm": 5.253132343292236, |
|
"learning_rate": 4.9783670141354776e-05, |
|
"loss": 0.042, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.3654217078730493, |
|
"grad_norm": 26.310007095336914, |
|
"learning_rate": 4.978176273719957e-05, |
|
"loss": 0.2762, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.3661230931088901, |
|
"grad_norm": 9.570679664611816, |
|
"learning_rate": 4.977984699787829e-05, |
|
"loss": 0.0898, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.36682447834473086, |
|
"grad_norm": 214.85060119628906, |
|
"learning_rate": 4.977792292403529e-05, |
|
"loss": 6.7239, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.36752586358057165, |
|
"grad_norm": 2.7513771057128906, |
|
"learning_rate": 4.977599051631773e-05, |
|
"loss": 0.0228, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.3682272488164124, |
|
"grad_norm": 33.629356384277344, |
|
"learning_rate": 4.977404977537554e-05, |
|
"loss": 0.4374, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.3689286340522532, |
|
"grad_norm": 3.8864808082580566, |
|
"learning_rate": 4.97721007018615e-05, |
|
"loss": 0.0308, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.36963001928809397, |
|
"grad_norm": 1.0654228925704956, |
|
"learning_rate": 4.9770143296431144e-05, |
|
"loss": 0.0077, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.37033140452393476, |
|
"grad_norm": 131.74990844726562, |
|
"learning_rate": 4.9768177559742844e-05, |
|
"loss": 4.0998, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.37103278975977555, |
|
"grad_norm": 0.10420957952737808, |
|
"learning_rate": 4.976620349245776e-05, |
|
"loss": 0.0006, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.37173417499561634, |
|
"grad_norm": 72.78388977050781, |
|
"learning_rate": 4.976422109523985e-05, |
|
"loss": 3.8242, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.37243556023145713, |
|
"grad_norm": 69.0342025756836, |
|
"learning_rate": 4.976223036875588e-05, |
|
"loss": 1.9872, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.3731369454672979, |
|
"grad_norm": 0.43323075771331787, |
|
"learning_rate": 4.9760231313675426e-05, |
|
"loss": 0.0028, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.3738383307031387, |
|
"grad_norm": 1.0747731924057007, |
|
"learning_rate": 4.975822393067084e-05, |
|
"loss": 0.0061, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.3745397159389795, |
|
"grad_norm": 69.71436309814453, |
|
"learning_rate": 4.9756208220417304e-05, |
|
"loss": 1.8796, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.3752411011748203, |
|
"grad_norm": 127.05543518066406, |
|
"learning_rate": 4.975418418359278e-05, |
|
"loss": 2.8441, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.3759424864106611, |
|
"grad_norm": 0.9959923624992371, |
|
"learning_rate": 4.975215182087804e-05, |
|
"loss": 0.0065, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.3766438716465018, |
|
"grad_norm": 2.1746952533721924, |
|
"learning_rate": 4.9750111132956654e-05, |
|
"loss": 0.0156, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.3773452568823426, |
|
"grad_norm": 92.57381439208984, |
|
"learning_rate": 4.9748062120514996e-05, |
|
"loss": 1.4085, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.3780466421181834, |
|
"grad_norm": 75.45511627197266, |
|
"learning_rate": 4.974600478424223e-05, |
|
"loss": 0.9909, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.3787480273540242, |
|
"grad_norm": 25.877460479736328, |
|
"learning_rate": 4.974393912483034e-05, |
|
"loss": 0.3578, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.379449412589865, |
|
"grad_norm": 12.262901306152344, |
|
"learning_rate": 4.9741865142974076e-05, |
|
"loss": 0.1338, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.38015079782570577, |
|
"grad_norm": 10.54548168182373, |
|
"learning_rate": 4.973978283937102e-05, |
|
"loss": 0.2111, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.38085218306154656, |
|
"grad_norm": 35.909976959228516, |
|
"learning_rate": 4.973769221472153e-05, |
|
"loss": 0.5549, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.38155356829738735, |
|
"grad_norm": 13.012696266174316, |
|
"learning_rate": 4.973559326972878e-05, |
|
"loss": 0.1784, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.38225495353322814, |
|
"grad_norm": 16.022974014282227, |
|
"learning_rate": 4.973348600509874e-05, |
|
"loss": 0.2235, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.3829563387690689, |
|
"grad_norm": 22.86570167541504, |
|
"learning_rate": 4.973137042154016e-05, |
|
"loss": 0.3561, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.3836577240049097, |
|
"grad_norm": 13.543822288513184, |
|
"learning_rate": 4.972924651976461e-05, |
|
"loss": 0.3077, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.3843591092407505, |
|
"grad_norm": 17.42881965637207, |
|
"learning_rate": 4.972711430048644e-05, |
|
"loss": 0.2779, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.38506049447659124, |
|
"grad_norm": 19.7313175201416, |
|
"learning_rate": 4.972497376442282e-05, |
|
"loss": 0.2582, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.38576187971243203, |
|
"grad_norm": 12.354894638061523, |
|
"learning_rate": 4.9722824912293695e-05, |
|
"loss": 0.2374, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.3864632649482728, |
|
"grad_norm": 17.16146469116211, |
|
"learning_rate": 4.972066774482182e-05, |
|
"loss": 0.2085, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.3871646501841136, |
|
"grad_norm": 49.288734436035156, |
|
"learning_rate": 4.971850226273274e-05, |
|
"loss": 0.6211, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.3878660354199544, |
|
"grad_norm": 4.870689392089844, |
|
"learning_rate": 4.97163284667548e-05, |
|
"loss": 0.0413, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.3885674206557952, |
|
"grad_norm": 4.268405437469482, |
|
"learning_rate": 4.971414635761914e-05, |
|
"loss": 0.0354, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.389268805891636, |
|
"grad_norm": 2.800025701522827, |
|
"learning_rate": 4.971195593605969e-05, |
|
"loss": 0.0215, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.3899701911274768, |
|
"grad_norm": 136.0956573486328, |
|
"learning_rate": 4.9709757202813195e-05, |
|
"loss": 3.3931, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.39067157636331756, |
|
"grad_norm": 68.56623840332031, |
|
"learning_rate": 4.970755015861918e-05, |
|
"loss": 1.8678, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.39137296159915835, |
|
"grad_norm": 68.25215911865234, |
|
"learning_rate": 4.970533480421996e-05, |
|
"loss": 1.9984, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.39207434683499914, |
|
"grad_norm": 69.02233123779297, |
|
"learning_rate": 4.9703111140360656e-05, |
|
"loss": 2.0662, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.39277573207083993, |
|
"grad_norm": 61.60824203491211, |
|
"learning_rate": 4.970087916778918e-05, |
|
"loss": 1.5281, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.39347711730668067, |
|
"grad_norm": 48.50178146362305, |
|
"learning_rate": 4.9698638887256254e-05, |
|
"loss": 0.8115, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.39417850254252146, |
|
"grad_norm": 41.32649612426758, |
|
"learning_rate": 4.969639029951537e-05, |
|
"loss": 0.6374, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.39487988777836225, |
|
"grad_norm": 6.946910858154297, |
|
"learning_rate": 4.9694133405322815e-05, |
|
"loss": 0.2402, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.39558127301420304, |
|
"grad_norm": 22.8358097076416, |
|
"learning_rate": 4.9691868205437694e-05, |
|
"loss": 0.3883, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.39628265825004383, |
|
"grad_norm": 15.402759552001953, |
|
"learning_rate": 4.968959470062189e-05, |
|
"loss": 0.7708, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.3969840434858846, |
|
"grad_norm": 28.08781623840332, |
|
"learning_rate": 4.968731289164006e-05, |
|
"loss": 0.7703, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.3976854287217254, |
|
"grad_norm": 26.910886764526367, |
|
"learning_rate": 4.9685022779259696e-05, |
|
"loss": 0.6299, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.3983868139575662, |
|
"grad_norm": 24.341957092285156, |
|
"learning_rate": 4.968272436425105e-05, |
|
"loss": 0.5751, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.399088199193407, |
|
"grad_norm": 23.469263076782227, |
|
"learning_rate": 4.9680417647387186e-05, |
|
"loss": 0.4664, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.3997895844292478, |
|
"grad_norm": 15.347972869873047, |
|
"learning_rate": 4.967810262944394e-05, |
|
"loss": 0.2252, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.40049096966508857, |
|
"grad_norm": 5.6460280418396, |
|
"learning_rate": 4.967577931119996e-05, |
|
"loss": 0.2232, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.40119235490092936, |
|
"grad_norm": 4.2383856773376465, |
|
"learning_rate": 4.967344769343667e-05, |
|
"loss": 0.0396, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.4018937401367701, |
|
"grad_norm": 1.6932095289230347, |
|
"learning_rate": 4.96711077769383e-05, |
|
"loss": 0.0137, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.4025951253726109, |
|
"grad_norm": 0.41152113676071167, |
|
"learning_rate": 4.966875956249186e-05, |
|
"loss": 0.0029, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.4032965106084517, |
|
"grad_norm": 0.06895897537469864, |
|
"learning_rate": 4.9666403050887155e-05, |
|
"loss": 0.0004, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.40399789584429247, |
|
"grad_norm": 0.004564318805932999, |
|
"learning_rate": 4.9664038242916786e-05, |
|
"loss": 0.0, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.40469928108013326, |
|
"grad_norm": 0.0019707169849425554, |
|
"learning_rate": 4.966166513937613e-05, |
|
"loss": 0.0, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.40540066631597405, |
|
"grad_norm": 0.00037522497586905956, |
|
"learning_rate": 4.965928374106337e-05, |
|
"loss": 0.0, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.40610205155181484, |
|
"grad_norm": 5.6252967624459416e-05, |
|
"learning_rate": 4.965689404877946e-05, |
|
"loss": 0.0, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.4068034367876556, |
|
"grad_norm": 1.1128309779451229e-05, |
|
"learning_rate": 4.9654496063328174e-05, |
|
"loss": 0.0, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.4075048220234964, |
|
"grad_norm": 119.89021301269531, |
|
"learning_rate": 4.965208978551605e-05, |
|
"loss": 10.8344, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.4082062072593372, |
|
"grad_norm": 1.3851165931555443e-05, |
|
"learning_rate": 4.964967521615241e-05, |
|
"loss": 0.0, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.408907592495178, |
|
"grad_norm": 3.7249556044116616e-06, |
|
"learning_rate": 4.96472523560494e-05, |
|
"loss": 0.0, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.4096089777310188, |
|
"grad_norm": 2.917176971095614e-06, |
|
"learning_rate": 4.9644821206021906e-05, |
|
"loss": 0.0, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.4103103629668595, |
|
"grad_norm": 3.3871406230900902e-06, |
|
"learning_rate": 4.964238176688765e-05, |
|
"loss": 0.0, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.4110117482027003, |
|
"grad_norm": 2.260056135128252e-06, |
|
"learning_rate": 4.9639934039467106e-05, |
|
"loss": 0.0, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.4117131334385411, |
|
"grad_norm": 1.5038928040667088e-06, |
|
"learning_rate": 4.963747802458355e-05, |
|
"loss": 0.0, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.4124145186743819, |
|
"grad_norm": 59.293521881103516, |
|
"learning_rate": 4.963501372306306e-05, |
|
"loss": 5.3994, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.4131159039102227, |
|
"grad_norm": 59.763816833496094, |
|
"learning_rate": 4.963254113573446e-05, |
|
"loss": 5.6505, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.41381728914606347, |
|
"grad_norm": 132.7192840576172, |
|
"learning_rate": 4.963006026342941e-05, |
|
"loss": 9.1499, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.41451867438190426, |
|
"grad_norm": 0.00024285895051434636, |
|
"learning_rate": 4.962757110698232e-05, |
|
"loss": 0.0, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.41522005961774505, |
|
"grad_norm": 0.0009618580807000399, |
|
"learning_rate": 4.9625073667230404e-05, |
|
"loss": 0.0, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.41592144485358584, |
|
"grad_norm": 0.004245178773999214, |
|
"learning_rate": 4.962256794501365e-05, |
|
"loss": 0.0, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.41662283008942663, |
|
"grad_norm": 0.01505737192928791, |
|
"learning_rate": 4.962005394117486e-05, |
|
"loss": 0.0001, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.4173242153252674, |
|
"grad_norm": 0.01870640181005001, |
|
"learning_rate": 4.961753165655958e-05, |
|
"loss": 0.0001, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.4180256005611082, |
|
"grad_norm": 0.047609515488147736, |
|
"learning_rate": 4.961500109201618e-05, |
|
"loss": 0.0003, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.41872698579694895, |
|
"grad_norm": 0.15332278609275818, |
|
"learning_rate": 4.9612462248395775e-05, |
|
"loss": 0.001, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.41942837103278974, |
|
"grad_norm": 113.53760528564453, |
|
"learning_rate": 4.96099151265523e-05, |
|
"loss": 2.3303, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.4201297562686305, |
|
"grad_norm": 0.6282661557197571, |
|
"learning_rate": 4.9607359727342465e-05, |
|
"loss": 0.0044, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.4208311415044713, |
|
"grad_norm": 43.630615234375, |
|
"learning_rate": 4.9604796051625756e-05, |
|
"loss": 0.6733, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.4215325267403121, |
|
"grad_norm": 72.0724868774414, |
|
"learning_rate": 4.960222410026444e-05, |
|
"loss": 1.4145, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.4222339119761529, |
|
"grad_norm": 6.166507244110107, |
|
"learning_rate": 4.959964387412358e-05, |
|
"loss": 0.0608, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.4229352972119937, |
|
"grad_norm": 80.03913879394531, |
|
"learning_rate": 4.9597055374071024e-05, |
|
"loss": 1.2555, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.4236366824478345, |
|
"grad_norm": 17.456140518188477, |
|
"learning_rate": 4.959445860097738e-05, |
|
"loss": 0.238, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.42433806768367527, |
|
"grad_norm": 36.63374328613281, |
|
"learning_rate": 4.9591853555716064e-05, |
|
"loss": 0.8324, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.42503945291951606, |
|
"grad_norm": 16.27304458618164, |
|
"learning_rate": 4.9589240239163255e-05, |
|
"loss": 0.2125, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.42574083815535685, |
|
"grad_norm": 19.427776336669922, |
|
"learning_rate": 4.9586618652197934e-05, |
|
"loss": 0.2701, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.42644222339119764, |
|
"grad_norm": 8.516806602478027, |
|
"learning_rate": 4.958398879570185e-05, |
|
"loss": 0.3408, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.4271436086270384, |
|
"grad_norm": 1.857074499130249, |
|
"learning_rate": 4.9581350670559524e-05, |
|
"loss": 0.2039, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.42784499386287916, |
|
"grad_norm": 8.925201416015625, |
|
"learning_rate": 4.9578704277658286e-05, |
|
"loss": 0.09, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.42854637909871995, |
|
"grad_norm": 4.490631580352783, |
|
"learning_rate": 4.957604961788823e-05, |
|
"loss": 0.0384, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.42924776433456074, |
|
"grad_norm": 0.866288959980011, |
|
"learning_rate": 4.957338669214222e-05, |
|
"loss": 0.0063, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.42994914957040153, |
|
"grad_norm": 0.6116040349006653, |
|
"learning_rate": 4.9570715501315925e-05, |
|
"loss": 0.0035, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.4306505348062423, |
|
"grad_norm": 0.014350208453834057, |
|
"learning_rate": 4.956803604630777e-05, |
|
"loss": 0.0001, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.4313519200420831, |
|
"grad_norm": 0.004978867247700691, |
|
"learning_rate": 4.956534832801897e-05, |
|
"loss": 0.0, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.4320533052779239, |
|
"grad_norm": 0.0008859955123625696, |
|
"learning_rate": 4.9562652347353535e-05, |
|
"loss": 0.0, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.4327546905137647, |
|
"grad_norm": 0.0007402606424875557, |
|
"learning_rate": 4.955994810521822e-05, |
|
"loss": 0.0, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.4334560757496055, |
|
"grad_norm": 60.98609161376953, |
|
"learning_rate": 4.955723560252259e-05, |
|
"loss": 4.6234, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.4341574609854463, |
|
"grad_norm": 60.63393783569336, |
|
"learning_rate": 4.955451484017896e-05, |
|
"loss": 4.669, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.43485884622128707, |
|
"grad_norm": 0.00016736971156205982, |
|
"learning_rate": 4.955178581910246e-05, |
|
"loss": 0.0, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.4355602314571278, |
|
"grad_norm": 0.00026081904070451856, |
|
"learning_rate": 4.9549048540210963e-05, |
|
"loss": 0.0, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.4362616166929686, |
|
"grad_norm": 0.0002771130239125341, |
|
"learning_rate": 4.954630300442513e-05, |
|
"loss": 0.0, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.4369630019288094, |
|
"grad_norm": 0.00024890206987038255, |
|
"learning_rate": 4.954354921266841e-05, |
|
"loss": 0.0, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.43766438716465017, |
|
"grad_norm": 0.0003540183824952692, |
|
"learning_rate": 4.954078716586702e-05, |
|
"loss": 0.0, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.43836577240049096, |
|
"grad_norm": 61.1258659362793, |
|
"learning_rate": 4.9538016864949953e-05, |
|
"loss": 4.1063, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.43906715763633175, |
|
"grad_norm": 61.82120895385742, |
|
"learning_rate": 4.953523831084897e-05, |
|
"loss": 3.2575, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.43976854287217254, |
|
"grad_norm": 0.009341489523649216, |
|
"learning_rate": 4.9532451504498644e-05, |
|
"loss": 0.0001, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.44046992810801333, |
|
"grad_norm": 126.25260925292969, |
|
"learning_rate": 4.952965644683627e-05, |
|
"loss": 4.8894, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.4411713133438541, |
|
"grad_norm": 64.0586166381836, |
|
"learning_rate": 4.9526853138801954e-05, |
|
"loss": 2.0939, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.4418726985796949, |
|
"grad_norm": 66.46090698242188, |
|
"learning_rate": 4.952404158133857e-05, |
|
"loss": 2.1672, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.4425740838155357, |
|
"grad_norm": 8.158260345458984, |
|
"learning_rate": 4.952122177539177e-05, |
|
"loss": 0.089, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.4432754690513765, |
|
"grad_norm": 8.28303337097168, |
|
"learning_rate": 4.951839372190996e-05, |
|
"loss": 0.1792, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.4439768542872172, |
|
"grad_norm": 6.670780658721924, |
|
"learning_rate": 4.951555742184435e-05, |
|
"loss": 0.3495, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.444678239523058, |
|
"grad_norm": 18.7882022857666, |
|
"learning_rate": 4.95127128761489e-05, |
|
"loss": 0.2761, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.4453796247588988, |
|
"grad_norm": 3.4713053703308105, |
|
"learning_rate": 4.950986008578037e-05, |
|
"loss": 0.2429, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.4460810099947396, |
|
"grad_norm": 8.415278434753418, |
|
"learning_rate": 4.950699905169825e-05, |
|
"loss": 0.0883, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.4467823952305804, |
|
"grad_norm": 5.512305736541748, |
|
"learning_rate": 4.950412977486484e-05, |
|
"loss": 0.0515, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.4474837804664212, |
|
"grad_norm": 1.983485460281372, |
|
"learning_rate": 4.95012522562452e-05, |
|
"loss": 0.0162, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.44818516570226197, |
|
"grad_norm": 0.5145743489265442, |
|
"learning_rate": 4.949836649680717e-05, |
|
"loss": 0.0038, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.44888655093810276, |
|
"grad_norm": 0.0464518703520298, |
|
"learning_rate": 4.949547249752134e-05, |
|
"loss": 0.0003, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.44958793617394355, |
|
"grad_norm": 63.47880935668945, |
|
"learning_rate": 4.9492570259361094e-05, |
|
"loss": 2.606, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.45028932140978434, |
|
"grad_norm": 63.044219970703125, |
|
"learning_rate": 4.948965978330258e-05, |
|
"loss": 3.3714, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.45099070664562513, |
|
"grad_norm": 61.190128326416016, |
|
"learning_rate": 4.948674107032471e-05, |
|
"loss": 3.3129, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.4516920918814659, |
|
"grad_norm": 59.03745651245117, |
|
"learning_rate": 4.948381412140917e-05, |
|
"loss": 2.3587, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.4523934771173067, |
|
"grad_norm": 61.125858306884766, |
|
"learning_rate": 4.948087893754043e-05, |
|
"loss": 2.7351, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.45309486235314744, |
|
"grad_norm": 60.69954299926758, |
|
"learning_rate": 4.947793551970571e-05, |
|
"loss": 2.2391, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.45379624758898823, |
|
"grad_norm": 56.91267395019531, |
|
"learning_rate": 4.9474983868895e-05, |
|
"loss": 1.5786, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.454497632824829, |
|
"grad_norm": 0.8762818574905396, |
|
"learning_rate": 4.947202398610108e-05, |
|
"loss": 0.0067, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.4551990180606698, |
|
"grad_norm": 2.4974522590637207, |
|
"learning_rate": 4.9469055872319483e-05, |
|
"loss": 0.0224, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.4559004032965106, |
|
"grad_norm": 22.98019790649414, |
|
"learning_rate": 4.9466079528548505e-05, |
|
"loss": 0.3665, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.4566017885323514, |
|
"grad_norm": 16.03438949584961, |
|
"learning_rate": 4.9463094955789214e-05, |
|
"loss": 0.315, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.4573031737681922, |
|
"grad_norm": 8.860062599182129, |
|
"learning_rate": 4.946010215504546e-05, |
|
"loss": 0.1067, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.458004559004033, |
|
"grad_norm": 5.521644115447998, |
|
"learning_rate": 4.945710112732386e-05, |
|
"loss": 0.2473, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.45870594423987376, |
|
"grad_norm": 15.16481876373291, |
|
"learning_rate": 4.945409187363376e-05, |
|
"loss": 0.2257, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.45940732947571455, |
|
"grad_norm": 15.714698791503906, |
|
"learning_rate": 4.945107439498732e-05, |
|
"loss": 0.2359, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.46010871471155534, |
|
"grad_norm": 6.4135589599609375, |
|
"learning_rate": 4.944804869239944e-05, |
|
"loss": 0.2336, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.46081009994739613, |
|
"grad_norm": 9.600846290588379, |
|
"learning_rate": 4.94450147668878e-05, |
|
"loss": 0.1174, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.46151148518323687, |
|
"grad_norm": 7.477170467376709, |
|
"learning_rate": 4.944197261947283e-05, |
|
"loss": 0.0857, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.46221287041907766, |
|
"grad_norm": 2.0777835845947266, |
|
"learning_rate": 4.943892225117774e-05, |
|
"loss": 0.0196, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.46291425565491845, |
|
"grad_norm": 41.072837829589844, |
|
"learning_rate": 4.9435863663028495e-05, |
|
"loss": 0.7781, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.46361564089075924, |
|
"grad_norm": 0.5752171277999878, |
|
"learning_rate": 4.943279685605383e-05, |
|
"loss": 0.0045, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.46431702612660003, |
|
"grad_norm": 51.6488151550293, |
|
"learning_rate": 4.9429721831285246e-05, |
|
"loss": 1.1945, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.4650184113624408, |
|
"grad_norm": 0.11998272687196732, |
|
"learning_rate": 4.942663858975701e-05, |
|
"loss": 0.0009, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.4657197965982816, |
|
"grad_norm": 57.40029525756836, |
|
"learning_rate": 4.9423547132506135e-05, |
|
"loss": 1.7823, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.4664211818341224, |
|
"grad_norm": 0.07954677194356918, |
|
"learning_rate": 4.942044746057242e-05, |
|
"loss": 0.0006, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.4671225670699632, |
|
"grad_norm": 77.24893951416016, |
|
"learning_rate": 4.941733957499841e-05, |
|
"loss": 3.3749, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.467823952305804, |
|
"grad_norm": 57.04423522949219, |
|
"learning_rate": 4.9414223476829434e-05, |
|
"loss": 1.6771, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.46852533754164477, |
|
"grad_norm": 0.14917835593223572, |
|
"learning_rate": 4.941109916711355e-05, |
|
"loss": 0.0011, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.46922672277748556, |
|
"grad_norm": 47.98805236816406, |
|
"learning_rate": 4.940796664690161e-05, |
|
"loss": 1.0673, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.4699281080133263, |
|
"grad_norm": 63.10405349731445, |
|
"learning_rate": 4.940482591724721e-05, |
|
"loss": 1.7249, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.4706294932491671, |
|
"grad_norm": 47.01327133178711, |
|
"learning_rate": 4.940167697920671e-05, |
|
"loss": 0.9642, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.4713308784850079, |
|
"grad_norm": 4.804792881011963, |
|
"learning_rate": 4.9398519833839234e-05, |
|
"loss": 0.0514, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.47203226372084867, |
|
"grad_norm": 46.21573257446289, |
|
"learning_rate": 4.939535448220667e-05, |
|
"loss": 0.7498, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.47273364895668946, |
|
"grad_norm": 12.326519966125488, |
|
"learning_rate": 4.939218092537365e-05, |
|
"loss": 0.1734, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.47343503419253025, |
|
"grad_norm": 2.402872085571289, |
|
"learning_rate": 4.938899916440759e-05, |
|
"loss": 0.2545, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.47413641942837104, |
|
"grad_norm": 4.058804512023926, |
|
"learning_rate": 4.938580920037864e-05, |
|
"loss": 0.2573, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.4748378046642118, |
|
"grad_norm": 11.151541709899902, |
|
"learning_rate": 4.938261103435973e-05, |
|
"loss": 0.4285, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.4755391899000526, |
|
"grad_norm": 6.957059860229492, |
|
"learning_rate": 4.937940466742653e-05, |
|
"loss": 0.2347, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.4762405751358934, |
|
"grad_norm": 5.085255146026611, |
|
"learning_rate": 4.9376190100657495e-05, |
|
"loss": 0.2671, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.4769419603717342, |
|
"grad_norm": 9.267155647277832, |
|
"learning_rate": 4.93729673351338e-05, |
|
"loss": 0.1125, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.477643345607575, |
|
"grad_norm": 6.162266731262207, |
|
"learning_rate": 4.9369736371939426e-05, |
|
"loss": 0.0683, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.4783447308434157, |
|
"grad_norm": 3.6411261558532715, |
|
"learning_rate": 4.936649721216105e-05, |
|
"loss": 0.0356, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.4790461160792565, |
|
"grad_norm": 48.720794677734375, |
|
"learning_rate": 4.9363249856888165e-05, |
|
"loss": 1.0375, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.4797475013150973, |
|
"grad_norm": 0.3436586856842041, |
|
"learning_rate": 4.9359994307212994e-05, |
|
"loss": 0.0027, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.4804488865509381, |
|
"grad_norm": 52.97607421875, |
|
"learning_rate": 4.93567305642305e-05, |
|
"loss": 1.6276, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.4811502717867789, |
|
"grad_norm": 0.09510497748851776, |
|
"learning_rate": 4.935345862903844e-05, |
|
"loss": 0.0007, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.4818516570226197, |
|
"grad_norm": 56.231292724609375, |
|
"learning_rate": 4.93501785027373e-05, |
|
"loss": 2.0342, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.48255304225846046, |
|
"grad_norm": 0.21096964180469513, |
|
"learning_rate": 4.934689018643032e-05, |
|
"loss": 0.0015, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.48325442749430125, |
|
"grad_norm": 0.19893679022789001, |
|
"learning_rate": 4.93435936812235e-05, |
|
"loss": 0.0014, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.48395581273014204, |
|
"grad_norm": 51.99702072143555, |
|
"learning_rate": 4.9340288988225594e-05, |
|
"loss": 1.0089, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.48465719796598283, |
|
"grad_norm": 53.519405364990234, |
|
"learning_rate": 4.9336976108548126e-05, |
|
"loss": 1.7908, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.4853585832018236, |
|
"grad_norm": 106.84630584716797, |
|
"learning_rate": 4.9333655043305354e-05, |
|
"loss": 3.6845, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.4860599684376644, |
|
"grad_norm": 0.24259038269519806, |
|
"learning_rate": 4.9330325793614285e-05, |
|
"loss": 0.0019, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.48676135367350515, |
|
"grad_norm": 0.8419340252876282, |
|
"learning_rate": 4.93269883605947e-05, |
|
"loss": 0.0072, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.48746273890934594, |
|
"grad_norm": 1.204994559288025, |
|
"learning_rate": 4.932364274536912e-05, |
|
"loss": 0.0087, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.48816412414518673, |
|
"grad_norm": 1.1827445030212402, |
|
"learning_rate": 4.932028894906279e-05, |
|
"loss": 0.0104, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.4888655093810275, |
|
"grad_norm": 0.5520339608192444, |
|
"learning_rate": 4.931692697280378e-05, |
|
"loss": 0.0045, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.4895668946168683, |
|
"grad_norm": 0.9636574983596802, |
|
"learning_rate": 4.931355681772283e-05, |
|
"loss": 0.0079, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.4902682798527091, |
|
"grad_norm": 0.5019407868385315, |
|
"learning_rate": 4.931017848495348e-05, |
|
"loss": 0.004, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.4909696650885499, |
|
"grad_norm": 0.0757029727101326, |
|
"learning_rate": 4.9306791975632026e-05, |
|
"loss": 0.0006, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.4916710503243907, |
|
"grad_norm": 110.13203430175781, |
|
"learning_rate": 4.9303397290897465e-05, |
|
"loss": 4.3731, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.49237243556023147, |
|
"grad_norm": 0.03152499720454216, |
|
"learning_rate": 4.929999443189159e-05, |
|
"loss": 0.0002, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.49307382079607226, |
|
"grad_norm": 0.029290443286299706, |
|
"learning_rate": 4.929658339975893e-05, |
|
"loss": 0.0002, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.49377520603191305, |
|
"grad_norm": 0.015484875068068504, |
|
"learning_rate": 4.9293164195646754e-05, |
|
"loss": 0.0001, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.49447659126775384, |
|
"grad_norm": 53.61540222167969, |
|
"learning_rate": 4.928973682070509e-05, |
|
"loss": 2.7716, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.4951779765035946, |
|
"grad_norm": 55.3106575012207, |
|
"learning_rate": 4.9286301276086715e-05, |
|
"loss": 2.3966, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.49587936173943536, |
|
"grad_norm": 0.014001145958900452, |
|
"learning_rate": 4.928285756294714e-05, |
|
"loss": 0.0001, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.49658074697527615, |
|
"grad_norm": 0.026434065774083138, |
|
"learning_rate": 4.927940568244465e-05, |
|
"loss": 0.0002, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.49728213221111695, |
|
"grad_norm": 0.04528193548321724, |
|
"learning_rate": 4.927594563574023e-05, |
|
"loss": 0.0003, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.49798351744695774, |
|
"grad_norm": 0.0950818806886673, |
|
"learning_rate": 4.9272477423997666e-05, |
|
"loss": 0.0007, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.4986849026827985, |
|
"grad_norm": 0.1095154732465744, |
|
"learning_rate": 4.926900104838346e-05, |
|
"loss": 0.0008, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.4993862879186393, |
|
"grad_norm": 0.13107261061668396, |
|
"learning_rate": 4.9265516510066864e-05, |
|
"loss": 0.001, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.5000876731544801, |
|
"grad_norm": 53.58122634887695, |
|
"learning_rate": 4.926202381021988e-05, |
|
"loss": 1.8605, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.5007890583903208, |
|
"grad_norm": 48.86884689331055, |
|
"learning_rate": 4.925852295001725e-05, |
|
"loss": 1.3723, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.5014904436261617, |
|
"grad_norm": 0.40202492475509644, |
|
"learning_rate": 4.9255013930636465e-05, |
|
"loss": 0.0033, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.5021918288620024, |
|
"grad_norm": 43.420196533203125, |
|
"learning_rate": 4.925149675325775e-05, |
|
"loss": 0.9394, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.5028932140978433, |
|
"grad_norm": 56.66259002685547, |
|
"learning_rate": 4.9247971419064097e-05, |
|
"loss": 1.2755, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.503594599333684, |
|
"grad_norm": 3.452730894088745, |
|
"learning_rate": 4.9244437929241215e-05, |
|
"loss": 0.0353, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.5042959845695248, |
|
"grad_norm": 4.454288005828857, |
|
"learning_rate": 4.924089628497757e-05, |
|
"loss": 0.0486, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.5049973698053656, |
|
"grad_norm": 18.87127685546875, |
|
"learning_rate": 4.923734648746437e-05, |
|
"loss": 0.3667, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.5056987550412064, |
|
"grad_norm": 11.85842514038086, |
|
"learning_rate": 4.923378853789556e-05, |
|
"loss": 0.3075, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.5064001402770472, |
|
"grad_norm": 9.63955307006836, |
|
"learning_rate": 4.923022243746783e-05, |
|
"loss": 0.1304, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.507101525512888, |
|
"grad_norm": 9.134459495544434, |
|
"learning_rate": 4.922664818738063e-05, |
|
"loss": 0.1346, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.5078029107487287, |
|
"grad_norm": 2.0840797424316406, |
|
"learning_rate": 4.92230657888361e-05, |
|
"loss": 0.2203, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.5085042959845695, |
|
"grad_norm": 8.857693672180176, |
|
"learning_rate": 4.921947524303919e-05, |
|
"loss": 0.1232, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.5092056812204103, |
|
"grad_norm": 8.82702350616455, |
|
"learning_rate": 4.921587655119753e-05, |
|
"loss": 0.1143, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.5099070664562511, |
|
"grad_norm": 4.017044544219971, |
|
"learning_rate": 4.9212269714521505e-05, |
|
"loss": 0.0433, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.5106084516920919, |
|
"grad_norm": 1.7636412382125854, |
|
"learning_rate": 4.920865473422428e-05, |
|
"loss": 0.0171, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.5113098369279326, |
|
"grad_norm": 107.01025390625, |
|
"learning_rate": 4.920503161152171e-05, |
|
"loss": 2.4773, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.5120112221637735, |
|
"grad_norm": 0.35363465547561646, |
|
"learning_rate": 4.9201400347632404e-05, |
|
"loss": 0.0028, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.5127126073996142, |
|
"grad_norm": 0.19405458867549896, |
|
"learning_rate": 4.919776094377771e-05, |
|
"loss": 0.0015, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.5134139926354551, |
|
"grad_norm": 55.00040054321289, |
|
"learning_rate": 4.919411340118173e-05, |
|
"loss": 1.9537, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.5141153778712958, |
|
"grad_norm": 55.760963439941406, |
|
"learning_rate": 4.919045772107127e-05, |
|
"loss": 2.1316, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.5148167631071366, |
|
"grad_norm": 55.24003601074219, |
|
"learning_rate": 4.91867939046759e-05, |
|
"loss": 2.1231, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.5155181483429774, |
|
"grad_norm": 0.03399452567100525, |
|
"learning_rate": 4.918312195322792e-05, |
|
"loss": 0.0002, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.5162195335788181, |
|
"grad_norm": 204.8887176513672, |
|
"learning_rate": 4.917944186796235e-05, |
|
"loss": 8.6129, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.516920918814659, |
|
"grad_norm": 124.1363754272461, |
|
"learning_rate": 4.917575365011699e-05, |
|
"loss": 3.9954, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.5176223040504997, |
|
"grad_norm": 0.14609888195991516, |
|
"learning_rate": 4.9172057300932306e-05, |
|
"loss": 0.0011, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.5183236892863405, |
|
"grad_norm": 0.33696234226226807, |
|
"learning_rate": 4.9168352821651575e-05, |
|
"loss": 0.0027, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.5190250745221813, |
|
"grad_norm": 105.28889465332031, |
|
"learning_rate": 4.916464021352075e-05, |
|
"loss": 3.0367, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.5197264597580221, |
|
"grad_norm": 1.017066240310669, |
|
"learning_rate": 4.916091947778855e-05, |
|
"loss": 0.0088, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.5204278449938629, |
|
"grad_norm": 1.143968939781189, |
|
"learning_rate": 4.91571906157064e-05, |
|
"loss": 0.0102, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.5211292302297037, |
|
"grad_norm": 0.7457045912742615, |
|
"learning_rate": 4.91534536285285e-05, |
|
"loss": 0.0063, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.5218306154655444, |
|
"grad_norm": 0.700978696346283, |
|
"learning_rate": 4.9149708517511743e-05, |
|
"loss": 0.0058, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.5225320007013853, |
|
"grad_norm": 0.38463106751441956, |
|
"learning_rate": 4.914595528391578e-05, |
|
"loss": 0.0031, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.523233385937226, |
|
"grad_norm": 0.14777745306491852, |
|
"learning_rate": 4.914219392900297e-05, |
|
"loss": 0.0011, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.5239347711730669, |
|
"grad_norm": 0.07884679734706879, |
|
"learning_rate": 4.913842445403843e-05, |
|
"loss": 0.0006, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.5246361564089076, |
|
"grad_norm": 61.55155563354492, |
|
"learning_rate": 4.913464686028999e-05, |
|
"loss": 1.992, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.5253375416447483, |
|
"grad_norm": 0.027440281584858894, |
|
"learning_rate": 4.913086114902822e-05, |
|
"loss": 0.0002, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.5260389268805892, |
|
"grad_norm": 55.66230010986328, |
|
"learning_rate": 4.912706732152641e-05, |
|
"loss": 2.231, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.5267403121164299, |
|
"grad_norm": 54.453102111816406, |
|
"learning_rate": 4.912326537906059e-05, |
|
"loss": 2.0983, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.5274416973522708, |
|
"grad_norm": 0.0637647733092308, |
|
"learning_rate": 4.911945532290951e-05, |
|
"loss": 0.0005, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.5281430825881115, |
|
"grad_norm": 0.13056063652038574, |
|
"learning_rate": 4.9115637154354674e-05, |
|
"loss": 0.001, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.5288444678239523, |
|
"grad_norm": 52.72209167480469, |
|
"learning_rate": 4.911181087468027e-05, |
|
"loss": 1.5666, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.5295458530597931, |
|
"grad_norm": 48.385467529296875, |
|
"learning_rate": 4.9107976485173255e-05, |
|
"loss": 1.2941, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.5302472382956339, |
|
"grad_norm": 0.5080452561378479, |
|
"learning_rate": 4.91041339871233e-05, |
|
"loss": 0.0041, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.5309486235314747, |
|
"grad_norm": 80.44468688964844, |
|
"learning_rate": 4.910028338182279e-05, |
|
"loss": 1.6558, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.5316500087673155, |
|
"grad_norm": 1.59328031539917, |
|
"learning_rate": 4.909642467056685e-05, |
|
"loss": 0.0141, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.5323513940031562, |
|
"grad_norm": 2.678886651992798, |
|
"learning_rate": 4.9092557854653344e-05, |
|
"loss": 0.0255, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.533052779238997, |
|
"grad_norm": 2.742159366607666, |
|
"learning_rate": 4.908868293538282e-05, |
|
"loss": 0.0271, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.5337541644748378, |
|
"grad_norm": 2.275885820388794, |
|
"learning_rate": 4.908479991405861e-05, |
|
"loss": 0.0215, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.5344555497106785, |
|
"grad_norm": 1.4059484004974365, |
|
"learning_rate": 4.9080908791986724e-05, |
|
"loss": 0.0126, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.5351569349465194, |
|
"grad_norm": 0.7012129426002502, |
|
"learning_rate": 4.907700957047592e-05, |
|
"loss": 0.0058, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.5358583201823601, |
|
"grad_norm": 0.2713419497013092, |
|
"learning_rate": 4.9073102250837655e-05, |
|
"loss": 0.0021, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.536559705418201, |
|
"grad_norm": 0.13228341937065125, |
|
"learning_rate": 4.9069186834386145e-05, |
|
"loss": 0.001, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.5372610906540417, |
|
"grad_norm": 0.06449166685342789, |
|
"learning_rate": 4.90652633224383e-05, |
|
"loss": 0.0005, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.5379624758898826, |
|
"grad_norm": 0.017581632360816002, |
|
"learning_rate": 4.9061331716313783e-05, |
|
"loss": 0.0001, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.5386638611257233, |
|
"grad_norm": 0.016885504126548767, |
|
"learning_rate": 4.905739201733495e-05, |
|
"loss": 0.0001, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.5393652463615641, |
|
"grad_norm": 0.005497175268828869, |
|
"learning_rate": 4.9053444226826886e-05, |
|
"loss": 0.0, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.5400666315974049, |
|
"grad_norm": 58.73657989501953, |
|
"learning_rate": 4.90494883461174e-05, |
|
"loss": 3.1923, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.5407680168332457, |
|
"grad_norm": 0.0022041036281734705, |
|
"learning_rate": 4.9045524376537034e-05, |
|
"loss": 0.0, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.5414694020690864, |
|
"grad_norm": 52.654136657714844, |
|
"learning_rate": 4.9041552319419035e-05, |
|
"loss": 3.097, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.5421707873049272, |
|
"grad_norm": 0.0026611709035933018, |
|
"learning_rate": 4.903757217609938e-05, |
|
"loss": 0.0, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.542872172540768, |
|
"grad_norm": 0.0028983913362026215, |
|
"learning_rate": 4.9033583947916765e-05, |
|
"loss": 0.0, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.5435735577766088, |
|
"grad_norm": 105.84739685058594, |
|
"learning_rate": 4.902958763621258e-05, |
|
"loss": 5.9716, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.5442749430124496, |
|
"grad_norm": 54.44437789916992, |
|
"learning_rate": 4.902558324233098e-05, |
|
"loss": 2.7511, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.5449763282482903, |
|
"grad_norm": 0.009829580783843994, |
|
"learning_rate": 4.9021570767618794e-05, |
|
"loss": 0.0001, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.5456777134841312, |
|
"grad_norm": 53.416908264160156, |
|
"learning_rate": 4.90175502134256e-05, |
|
"loss": 2.337, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.5463790987199719, |
|
"grad_norm": 0.050978660583496094, |
|
"learning_rate": 4.9013521581103686e-05, |
|
"loss": 0.0004, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.5470804839558128, |
|
"grad_norm": 0.11917990446090698, |
|
"learning_rate": 4.900948487200804e-05, |
|
"loss": 0.0009, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.5477818691916535, |
|
"grad_norm": 48.541015625, |
|
"learning_rate": 4.900544008749639e-05, |
|
"loss": 1.2937, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.5484832544274943, |
|
"grad_norm": 0.3613051474094391, |
|
"learning_rate": 4.900138722892917e-05, |
|
"loss": 0.003, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.5491846396633351, |
|
"grad_norm": 35.976261138916016, |
|
"learning_rate": 4.899732629766953e-05, |
|
"loss": 0.7424, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.5498860248991758, |
|
"grad_norm": 1.3365657329559326, |
|
"learning_rate": 4.899325729508333e-05, |
|
"loss": 0.012, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.5505874101350167, |
|
"grad_norm": 2.2534306049346924, |
|
"learning_rate": 4.898918022253916e-05, |
|
"loss": 0.0228, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.5512887953708574, |
|
"grad_norm": 24.77885627746582, |
|
"learning_rate": 4.898509508140829e-05, |
|
"loss": 0.4601, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.5519901806066982, |
|
"grad_norm": 4.28820276260376, |
|
"learning_rate": 4.898100187306475e-05, |
|
"loss": 0.0482, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.552691565842539, |
|
"grad_norm": 4.493249416351318, |
|
"learning_rate": 4.897690059888524e-05, |
|
"loss": 0.0503, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.5533929510783798, |
|
"grad_norm": 3.269726037979126, |
|
"learning_rate": 4.897279126024923e-05, |
|
"loss": 0.0349, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.5540943363142206, |
|
"grad_norm": 3.0403542518615723, |
|
"learning_rate": 4.896867385853883e-05, |
|
"loss": 0.0316, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.5547957215500614, |
|
"grad_norm": 25.840961456298828, |
|
"learning_rate": 4.896454839513892e-05, |
|
"loss": 0.4405, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.5554971067859021, |
|
"grad_norm": 72.00930786132812, |
|
"learning_rate": 4.8960414871437055e-05, |
|
"loss": 1.4578, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.556198492021743, |
|
"grad_norm": 0.7836804389953613, |
|
"learning_rate": 4.895627328882352e-05, |
|
"loss": 0.0068, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.5568998772575837, |
|
"grad_norm": 0.7037481069564819, |
|
"learning_rate": 4.895212364869132e-05, |
|
"loss": 0.0061, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.5576012624934246, |
|
"grad_norm": 0.4775547981262207, |
|
"learning_rate": 4.894796595243614e-05, |
|
"loss": 0.004, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.5583026477292653, |
|
"grad_norm": 0.4537871778011322, |
|
"learning_rate": 4.894380020145639e-05, |
|
"loss": 0.0038, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.559004032965106, |
|
"grad_norm": 0.2530267536640167, |
|
"learning_rate": 4.8939626397153196e-05, |
|
"loss": 0.0021, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.5597054182009469, |
|
"grad_norm": 0.08755356073379517, |
|
"learning_rate": 4.8935444540930385e-05, |
|
"loss": 0.0007, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.5604068034367876, |
|
"grad_norm": 0.058386508375406265, |
|
"learning_rate": 4.8931254634194504e-05, |
|
"loss": 0.0004, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.5611081886726285, |
|
"grad_norm": 0.020555956289172173, |
|
"learning_rate": 4.892705667835479e-05, |
|
"loss": 0.0002, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.5618095739084692, |
|
"grad_norm": 0.023624928668141365, |
|
"learning_rate": 4.892285067482319e-05, |
|
"loss": 0.0002, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.56251095914431, |
|
"grad_norm": 54.471675872802734, |
|
"learning_rate": 4.891863662501437e-05, |
|
"loss": 2.8657, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.5632123443801508, |
|
"grad_norm": 0.005444019101560116, |
|
"learning_rate": 4.8914414530345684e-05, |
|
"loss": 0.0, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.5639137296159916, |
|
"grad_norm": 0.008992165327072144, |
|
"learning_rate": 4.891018439223722e-05, |
|
"loss": 0.0001, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.5646151148518324, |
|
"grad_norm": 0.0062455967999994755, |
|
"learning_rate": 4.890594621211174e-05, |
|
"loss": 0.0, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.5653165000876732, |
|
"grad_norm": 0.006425914354622364, |
|
"learning_rate": 4.8901699991394735e-05, |
|
"loss": 0.0, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.5660178853235139, |
|
"grad_norm": 0.008212440647184849, |
|
"learning_rate": 4.8897445731514384e-05, |
|
"loss": 0.0001, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.5667192705593547, |
|
"grad_norm": 0.004950184375047684, |
|
"learning_rate": 4.889318343390158e-05, |
|
"loss": 0.0, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.5674206557951955, |
|
"grad_norm": 0.0056209079921245575, |
|
"learning_rate": 4.8888913099989916e-05, |
|
"loss": 0.0, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.5681220410310363, |
|
"grad_norm": 0.0028246166184544563, |
|
"learning_rate": 4.888463473121568e-05, |
|
"loss": 0.0, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.5688234262668771, |
|
"grad_norm": 53.988868713378906, |
|
"learning_rate": 4.888034832901788e-05, |
|
"loss": 2.7359, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.5695248115027178, |
|
"grad_norm": 0.004621135536581278, |
|
"learning_rate": 4.8876053894838215e-05, |
|
"loss": 0.0, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.5702261967385587, |
|
"grad_norm": 0.011877781711518764, |
|
"learning_rate": 4.887175143012108e-05, |
|
"loss": 0.0001, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.5709275819743994, |
|
"grad_norm": 65.13773345947266, |
|
"learning_rate": 4.886744093631358e-05, |
|
"loss": 3.5609, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.5716289672102403, |
|
"grad_norm": 161.86300659179688, |
|
"learning_rate": 4.886312241486554e-05, |
|
"loss": 7.2583, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.572330352446081, |
|
"grad_norm": 0.032078277319669724, |
|
"learning_rate": 4.8858795867229425e-05, |
|
"loss": 0.0002, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.5730317376819218, |
|
"grad_norm": 113.32162475585938, |
|
"learning_rate": 4.8854461294860454e-05, |
|
"loss": 4.1658, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.5737331229177626, |
|
"grad_norm": 0.17656919360160828, |
|
"learning_rate": 4.885011869921654e-05, |
|
"loss": 0.0014, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.5744345081536034, |
|
"grad_norm": 0.2188340574502945, |
|
"learning_rate": 4.884576808175828e-05, |
|
"loss": 0.0018, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.5751358933894442, |
|
"grad_norm": 0.38712555170059204, |
|
"learning_rate": 4.884140944394896e-05, |
|
"loss": 0.0033, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.5758372786252849, |
|
"grad_norm": 0.7210093140602112, |
|
"learning_rate": 4.883704278725458e-05, |
|
"loss": 0.0062, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.5765386638611257, |
|
"grad_norm": 0.7694672346115112, |
|
"learning_rate": 4.883266811314385e-05, |
|
"loss": 0.0069, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.5772400490969665, |
|
"grad_norm": 0.6353862881660461, |
|
"learning_rate": 4.8828285423088146e-05, |
|
"loss": 0.0056, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.5779414343328073, |
|
"grad_norm": 87.86915588378906, |
|
"learning_rate": 4.8823894718561545e-05, |
|
"loss": 2.1381, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.578642819568648, |
|
"grad_norm": 0.42429307103157043, |
|
"learning_rate": 4.8819496001040846e-05, |
|
"loss": 0.0036, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.5793442048044889, |
|
"grad_norm": 0.40686339139938354, |
|
"learning_rate": 4.881508927200552e-05, |
|
"loss": 0.0034, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.5800455900403296, |
|
"grad_norm": 0.27934542298316956, |
|
"learning_rate": 4.881067453293774e-05, |
|
"loss": 0.0022, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.5807469752761705, |
|
"grad_norm": 44.31092834472656, |
|
"learning_rate": 4.8806251785322355e-05, |
|
"loss": 1.1003, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.5814483605120112, |
|
"grad_norm": 0.17587609589099884, |
|
"learning_rate": 4.880182103064695e-05, |
|
"loss": 0.0014, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.582149745747852, |
|
"grad_norm": 0.2651273012161255, |
|
"learning_rate": 4.879738227040176e-05, |
|
"loss": 0.0021, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.5828511309836928, |
|
"grad_norm": 149.13035583496094, |
|
"learning_rate": 4.879293550607974e-05, |
|
"loss": 2.9684, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.5835525162195335, |
|
"grad_norm": 0.17069180309772491, |
|
"learning_rate": 4.878848073917651e-05, |
|
"loss": 0.0014, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.5842539014553744, |
|
"grad_norm": 0.24799254536628723, |
|
"learning_rate": 4.878401797119042e-05, |
|
"loss": 0.002, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.5849552866912151, |
|
"grad_norm": 0.15171018242835999, |
|
"learning_rate": 4.8779547203622477e-05, |
|
"loss": 0.0012, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.585656671927056, |
|
"grad_norm": 45.8453254699707, |
|
"learning_rate": 4.877506843797639e-05, |
|
"loss": 1.2116, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.5863580571628967, |
|
"grad_norm": 70.00684356689453, |
|
"learning_rate": 4.877058167575857e-05, |
|
"loss": 2.1496, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.5870594423987375, |
|
"grad_norm": 43.86625671386719, |
|
"learning_rate": 4.876608691847809e-05, |
|
"loss": 1.077, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.5877608276345783, |
|
"grad_norm": 39.75180435180664, |
|
"learning_rate": 4.876158416764675e-05, |
|
"loss": 0.8959, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.5884622128704191, |
|
"grad_norm": 78.85111999511719, |
|
"learning_rate": 4.8757073424779e-05, |
|
"loss": 1.609, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.5891635981062598, |
|
"grad_norm": 2.562579393386841, |
|
"learning_rate": 4.8752554691392004e-05, |
|
"loss": 0.0262, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.5898649833421007, |
|
"grad_norm": 3.174607276916504, |
|
"learning_rate": 4.8748027969005606e-05, |
|
"loss": 0.0331, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.5905663685779414, |
|
"grad_norm": 24.42951202392578, |
|
"learning_rate": 4.874349325914232e-05, |
|
"loss": 0.494, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.5912677538137823, |
|
"grad_norm": 5.76008415222168, |
|
"learning_rate": 4.87389505633274e-05, |
|
"loss": 0.069, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.591969139049623, |
|
"grad_norm": 16.484464645385742, |
|
"learning_rate": 4.873439988308871e-05, |
|
"loss": 0.3677, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.5926705242854637, |
|
"grad_norm": 8.441133499145508, |
|
"learning_rate": 4.872984121995686e-05, |
|
"loss": 0.1122, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.5933719095213046, |
|
"grad_norm": 7.171937942504883, |
|
"learning_rate": 4.8725274575465106e-05, |
|
"loss": 0.093, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.5940732947571453, |
|
"grad_norm": 7.630092620849609, |
|
"learning_rate": 4.872069995114943e-05, |
|
"loss": 0.0997, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.5947746799929862, |
|
"grad_norm": 5.521340370178223, |
|
"learning_rate": 4.871611734854845e-05, |
|
"loss": 0.065, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.5954760652288269, |
|
"grad_norm": 40.95110321044922, |
|
"learning_rate": 4.871152676920351e-05, |
|
"loss": 0.6631, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.5961774504646677, |
|
"grad_norm": 2.602055311203003, |
|
"learning_rate": 4.8706928214658596e-05, |
|
"loss": 0.0269, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.5968788357005085, |
|
"grad_norm": 30.752164840698242, |
|
"learning_rate": 4.8702321686460424e-05, |
|
"loss": 0.5944, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.5975802209363493, |
|
"grad_norm": 34.97576904296875, |
|
"learning_rate": 4.8697707186158345e-05, |
|
"loss": 0.7216, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.5982816061721901, |
|
"grad_norm": 1.3377373218536377, |
|
"learning_rate": 4.869308471530443e-05, |
|
"loss": 0.0126, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.5989829914080309, |
|
"grad_norm": 33.73777389526367, |
|
"learning_rate": 4.868845427545341e-05, |
|
"loss": 0.6655, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.5996843766438716, |
|
"grad_norm": 27.722604751586914, |
|
"learning_rate": 4.868381586816268e-05, |
|
"loss": 0.5284, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.6003857618797125, |
|
"grad_norm": 39.0617561340332, |
|
"learning_rate": 4.867916949499236e-05, |
|
"loss": 0.9141, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.6010871471155532, |
|
"grad_norm": 25.998207092285156, |
|
"learning_rate": 4.8674515157505205e-05, |
|
"loss": 0.4843, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.601788532351394, |
|
"grad_norm": 70.33052062988281, |
|
"learning_rate": 4.8669852857266686e-05, |
|
"loss": 1.1847, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.6024899175872348, |
|
"grad_norm": 2.859663724899292, |
|
"learning_rate": 4.866518259584492e-05, |
|
"loss": 0.1724, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.6031913028230755, |
|
"grad_norm": 7.6815185546875, |
|
"learning_rate": 4.866050437481072e-05, |
|
"loss": 0.2053, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.6038926880589164, |
|
"grad_norm": 15.574623107910156, |
|
"learning_rate": 4.8655818195737576e-05, |
|
"loss": 0.2786, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.6045940732947571, |
|
"grad_norm": 17.501440048217773, |
|
"learning_rate": 4.865112406020164e-05, |
|
"loss": 0.3404, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.605295458530598, |
|
"grad_norm": 18.48247718811035, |
|
"learning_rate": 4.8646421969781757e-05, |
|
"loss": 0.3676, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.6059968437664387, |
|
"grad_norm": 15.630892753601074, |
|
"learning_rate": 4.864171192605944e-05, |
|
"loss": 0.2856, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.6066982290022795, |
|
"grad_norm": 15.12441349029541, |
|
"learning_rate": 4.863699393061888e-05, |
|
"loss": 0.2728, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.6073996142381203, |
|
"grad_norm": 10.427019119262695, |
|
"learning_rate": 4.863226798504694e-05, |
|
"loss": 0.308, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.6081009994739611, |
|
"grad_norm": 10.065474510192871, |
|
"learning_rate": 4.8627534090933156e-05, |
|
"loss": 0.1617, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.6088023847098019, |
|
"grad_norm": 11.481346130371094, |
|
"learning_rate": 4.862279224986973e-05, |
|
"loss": 0.2136, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.6095037699456426, |
|
"grad_norm": 5.78438663482666, |
|
"learning_rate": 4.8618042463451566e-05, |
|
"loss": 0.0726, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.6102051551814834, |
|
"grad_norm": 39.987937927246094, |
|
"learning_rate": 4.8613284733276206e-05, |
|
"loss": 0.8901, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.6109065404173242, |
|
"grad_norm": 3.24749755859375, |
|
"learning_rate": 4.860851906094388e-05, |
|
"loss": 0.0369, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.611607925653165, |
|
"grad_norm": 2.4093921184539795, |
|
"learning_rate": 4.860374544805749e-05, |
|
"loss": 0.0254, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.6123093108890058, |
|
"grad_norm": 0.970287024974823, |
|
"learning_rate": 4.85989638962226e-05, |
|
"loss": 0.0092, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.6130106961248466, |
|
"grad_norm": 0.4123772084712982, |
|
"learning_rate": 4.859417440704746e-05, |
|
"loss": 0.0036, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.6137120813606873, |
|
"grad_norm": 0.14014503359794617, |
|
"learning_rate": 4.8589376982142977e-05, |
|
"loss": 0.0011, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.6144134665965282, |
|
"grad_norm": 0.05853430926799774, |
|
"learning_rate": 4.858457162312272e-05, |
|
"loss": 0.0005, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.6151148518323689, |
|
"grad_norm": 104.4763412475586, |
|
"learning_rate": 4.857975833160295e-05, |
|
"loss": 4.5228, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.6158162370682098, |
|
"grad_norm": 60.22207260131836, |
|
"learning_rate": 4.857493710920258e-05, |
|
"loss": 3.2127, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.6165176223040505, |
|
"grad_norm": 0.014606939628720284, |
|
"learning_rate": 4.8570107957543185e-05, |
|
"loss": 0.0001, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.6172190075398913, |
|
"grad_norm": 52.49748229980469, |
|
"learning_rate": 4.856527087824902e-05, |
|
"loss": 2.5181, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.6179203927757321, |
|
"grad_norm": 60.33863067626953, |
|
"learning_rate": 4.8560425872947004e-05, |
|
"loss": 3.4192, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.6186217780115728, |
|
"grad_norm": 0.017534319311380386, |
|
"learning_rate": 4.8555572943266715e-05, |
|
"loss": 0.0001, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.6193231632474137, |
|
"grad_norm": 0.03253171965479851, |
|
"learning_rate": 4.85507120908404e-05, |
|
"loss": 0.0002, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.6200245484832544, |
|
"grad_norm": 100.64522552490234, |
|
"learning_rate": 4.8545843317302983e-05, |
|
"loss": 3.6464, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.6207259337190952, |
|
"grad_norm": 52.276100158691406, |
|
"learning_rate": 4.854096662429203e-05, |
|
"loss": 1.5974, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.621427318954936, |
|
"grad_norm": 0.26515480875968933, |
|
"learning_rate": 4.853608201344778e-05, |
|
"loss": 0.0022, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.6221287041907768, |
|
"grad_norm": 0.4160686731338501, |
|
"learning_rate": 4.853118948641314e-05, |
|
"loss": 0.0036, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.6228300894266176, |
|
"grad_norm": 40.08952331542969, |
|
"learning_rate": 4.852628904483367e-05, |
|
"loss": 0.9215, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.6235314746624584, |
|
"grad_norm": 35.90172576904297, |
|
"learning_rate": 4.852138069035761e-05, |
|
"loss": 0.7604, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.6242328598982991, |
|
"grad_norm": 2.5494091510772705, |
|
"learning_rate": 4.851646442463585e-05, |
|
"loss": 0.0269, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.62493424513414, |
|
"grad_norm": 4.3015456199646, |
|
"learning_rate": 4.851154024932193e-05, |
|
"loss": 0.0495, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.6256356303699807, |
|
"grad_norm": 3.570831537246704, |
|
"learning_rate": 4.8506608166072054e-05, |
|
"loss": 0.0385, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.6263370156058214, |
|
"grad_norm": 3.575366735458374, |
|
"learning_rate": 4.8501668176545115e-05, |
|
"loss": 0.0397, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.6270384008416623, |
|
"grad_norm": 2.7419331073760986, |
|
"learning_rate": 4.849672028240262e-05, |
|
"loss": 0.0288, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.627739786077503, |
|
"grad_norm": 41.37546157836914, |
|
"learning_rate": 4.849176448530878e-05, |
|
"loss": 0.8202, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.6284411713133439, |
|
"grad_norm": 34.27451705932617, |
|
"learning_rate": 4.8486800786930415e-05, |
|
"loss": 0.7397, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.6291425565491846, |
|
"grad_norm": 1.0924946069717407, |
|
"learning_rate": 4.848182918893705e-05, |
|
"loss": 0.0104, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.6298439417850255, |
|
"grad_norm": 0.9149582386016846, |
|
"learning_rate": 4.8476849693000836e-05, |
|
"loss": 0.0085, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.6305453270208662, |
|
"grad_norm": 0.4877033829689026, |
|
"learning_rate": 4.8471862300796594e-05, |
|
"loss": 0.0043, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.631246712256707, |
|
"grad_norm": 50.69717788696289, |
|
"learning_rate": 4.846686701400179e-05, |
|
"loss": 1.3647, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.6319480974925478, |
|
"grad_norm": 45.04077911376953, |
|
"learning_rate": 4.846186383429657e-05, |
|
"loss": 1.366, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.6326494827283886, |
|
"grad_norm": 47.341819763183594, |
|
"learning_rate": 4.8456852763363706e-05, |
|
"loss": 1.2603, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.6333508679642293, |
|
"grad_norm": 44.485530853271484, |
|
"learning_rate": 4.8451833802888625e-05, |
|
"loss": 1.3248, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.6340522532000702, |
|
"grad_norm": 0.46667349338531494, |
|
"learning_rate": 4.844680695455943e-05, |
|
"loss": 0.0041, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.6347536384359109, |
|
"grad_norm": 0.7628422379493713, |
|
"learning_rate": 4.844177222006687e-05, |
|
"loss": 0.0067, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.6354550236717517, |
|
"grad_norm": 0.7451947927474976, |
|
"learning_rate": 4.8436729601104327e-05, |
|
"loss": 0.0069, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.6361564089075925, |
|
"grad_norm": 0.6644980907440186, |
|
"learning_rate": 4.843167909936786e-05, |
|
"loss": 0.006, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.6368577941434332, |
|
"grad_norm": 41.53312683105469, |
|
"learning_rate": 4.8426620716556164e-05, |
|
"loss": 1.1098, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.6375591793792741, |
|
"grad_norm": 0.5270825624465942, |
|
"learning_rate": 4.8421554454370596e-05, |
|
"loss": 0.0046, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.6382605646151148, |
|
"grad_norm": 90.92555236816406, |
|
"learning_rate": 4.841648031451514e-05, |
|
"loss": 2.5683, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.6389619498509557, |
|
"grad_norm": 46.597808837890625, |
|
"learning_rate": 4.841139829869646e-05, |
|
"loss": 1.4381, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.6396633350867964, |
|
"grad_norm": 0.9347837567329407, |
|
"learning_rate": 4.8406308408623865e-05, |
|
"loss": 0.0083, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.6403647203226372, |
|
"grad_norm": 0.8483829498291016, |
|
"learning_rate": 4.840121064600928e-05, |
|
"loss": 0.0077, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.641066105558478, |
|
"grad_norm": 37.393951416015625, |
|
"learning_rate": 4.8396105012567315e-05, |
|
"loss": 0.8573, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.6417674907943188, |
|
"grad_norm": 69.15557098388672, |
|
"learning_rate": 4.839099151001521e-05, |
|
"loss": 1.491, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.6424688760301596, |
|
"grad_norm": 1.3047503232955933, |
|
"learning_rate": 4.838587014007284e-05, |
|
"loss": 0.0126, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.6431702612660003, |
|
"grad_norm": 1.462365746498108, |
|
"learning_rate": 4.838074090446276e-05, |
|
"loss": 0.0144, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 0.6438716465018411, |
|
"grad_norm": 1.4955204725265503, |
|
"learning_rate": 4.8375603804910154e-05, |
|
"loss": 0.0146, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.6445730317376819, |
|
"grad_norm": 1.0593093633651733, |
|
"learning_rate": 4.837045884314283e-05, |
|
"loss": 0.0101, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 0.6452744169735227, |
|
"grad_norm": 0.5433834791183472, |
|
"learning_rate": 4.836530602089127e-05, |
|
"loss": 0.0049, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.6459758022093635, |
|
"grad_norm": 44.653175354003906, |
|
"learning_rate": 4.836014533988857e-05, |
|
"loss": 1.2462, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.6466771874452043, |
|
"grad_norm": 0.15624181926250458, |
|
"learning_rate": 4.835497680187052e-05, |
|
"loss": 0.0013, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 0.647378572681045, |
|
"grad_norm": 0.13333559036254883, |
|
"learning_rate": 4.8349800408575506e-05, |
|
"loss": 0.0011, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 0.6480799579168859, |
|
"grad_norm": 0.08697400987148285, |
|
"learning_rate": 4.834461616174456e-05, |
|
"loss": 0.0007, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.6487813431527266, |
|
"grad_norm": 49.62416076660156, |
|
"learning_rate": 4.8339424063121375e-05, |
|
"loss": 2.1264, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.6494827283885675, |
|
"grad_norm": 47.912906646728516, |
|
"learning_rate": 4.833422411445228e-05, |
|
"loss": 1.8048, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 0.6501841136244082, |
|
"grad_norm": 0.0816919356584549, |
|
"learning_rate": 4.832901631748622e-05, |
|
"loss": 0.0007, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.650885498860249, |
|
"grad_norm": 49.09192657470703, |
|
"learning_rate": 4.8323800673974826e-05, |
|
"loss": 1.7493, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 0.6515868840960898, |
|
"grad_norm": 98.90199279785156, |
|
"learning_rate": 4.8318577185672325e-05, |
|
"loss": 3.7461, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 0.6522882693319305, |
|
"grad_norm": 0.22995713353157043, |
|
"learning_rate": 4.831334585433561e-05, |
|
"loss": 0.002, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.6529896545677714, |
|
"grad_norm": 0.4566737413406372, |
|
"learning_rate": 4.830810668172419e-05, |
|
"loss": 0.0041, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 0.6536910398036121, |
|
"grad_norm": 0.5328819155693054, |
|
"learning_rate": 4.830285966960023e-05, |
|
"loss": 0.0048, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 0.6543924250394529, |
|
"grad_norm": 0.6611910462379456, |
|
"learning_rate": 4.829760481972852e-05, |
|
"loss": 0.0061, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.6550938102752937, |
|
"grad_norm": 76.24475860595703, |
|
"learning_rate": 4.82923421338765e-05, |
|
"loss": 1.7814, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 0.6557951955111345, |
|
"grad_norm": 0.7961150407791138, |
|
"learning_rate": 4.828707161381423e-05, |
|
"loss": 0.0075, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.6564965807469753, |
|
"grad_norm": 33.920326232910156, |
|
"learning_rate": 4.8281793261314404e-05, |
|
"loss": 0.7362, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.6571979659828161, |
|
"grad_norm": 0.9916706681251526, |
|
"learning_rate": 4.827650707815236e-05, |
|
"loss": 0.0096, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 0.6578993512186568, |
|
"grad_norm": 0.7823458909988403, |
|
"learning_rate": 4.827121306610608e-05, |
|
"loss": 0.0074, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 0.6586007364544977, |
|
"grad_norm": 0.8576016426086426, |
|
"learning_rate": 4.8265911226956154e-05, |
|
"loss": 0.008, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.6593021216903384, |
|
"grad_norm": 0.5242276787757874, |
|
"learning_rate": 4.826060156248582e-05, |
|
"loss": 0.0047, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.6600035069261792, |
|
"grad_norm": 0.31952106952667236, |
|
"learning_rate": 4.825528407448096e-05, |
|
"loss": 0.0028, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 0.66070489216202, |
|
"grad_norm": 0.17970554530620575, |
|
"learning_rate": 4.824995876473004e-05, |
|
"loss": 0.0015, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 0.6614062773978607, |
|
"grad_norm": 0.10021814703941345, |
|
"learning_rate": 4.824462563502422e-05, |
|
"loss": 0.0008, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 0.6621076626337016, |
|
"grad_norm": 146.59963989257812, |
|
"learning_rate": 4.823928468715724e-05, |
|
"loss": 5.6442, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 0.6628090478695423, |
|
"grad_norm": 0.060028087347745895, |
|
"learning_rate": 4.82339359229255e-05, |
|
"loss": 0.0005, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.6635104331053832, |
|
"grad_norm": 0.05350363254547119, |
|
"learning_rate": 4.822857934412802e-05, |
|
"loss": 0.0004, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 0.6642118183412239, |
|
"grad_norm": 197.06178283691406, |
|
"learning_rate": 4.822321495256644e-05, |
|
"loss": 8.2989, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 0.6649132035770647, |
|
"grad_norm": 0.047274697571992874, |
|
"learning_rate": 4.821784275004503e-05, |
|
"loss": 0.0004, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 0.6656145888129055, |
|
"grad_norm": 0.037716954946517944, |
|
"learning_rate": 4.82124627383707e-05, |
|
"loss": 0.0003, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 0.6663159740487463, |
|
"grad_norm": 48.574623107910156, |
|
"learning_rate": 4.8207074919352976e-05, |
|
"loss": 1.7823, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.667017359284587, |
|
"grad_norm": 0.06911084055900574, |
|
"learning_rate": 4.8201679294804e-05, |
|
"loss": 0.0006, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 0.6677187445204279, |
|
"grad_norm": 0.07155507057905197, |
|
"learning_rate": 4.8196275866538574e-05, |
|
"loss": 0.0006, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 0.6684201297562686, |
|
"grad_norm": 0.10516396909952164, |
|
"learning_rate": 4.8190864636374086e-05, |
|
"loss": 0.0009, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 0.6691215149921094, |
|
"grad_norm": 46.50364685058594, |
|
"learning_rate": 4.818544560613056e-05, |
|
"loss": 1.4989, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 0.6698229002279502, |
|
"grad_norm": 0.2198321670293808, |
|
"learning_rate": 4.8180018777630664e-05, |
|
"loss": 0.0019, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.670524285463791, |
|
"grad_norm": 87.00233459472656, |
|
"learning_rate": 4.8174584152699665e-05, |
|
"loss": 2.3952, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 0.6712256706996318, |
|
"grad_norm": 39.932960510253906, |
|
"learning_rate": 4.816914173316545e-05, |
|
"loss": 1.0167, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 0.6719270559354725, |
|
"grad_norm": 0.8517984747886658, |
|
"learning_rate": 4.816369152085855e-05, |
|
"loss": 0.0082, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.6726284411713134, |
|
"grad_norm": 1.0399830341339111, |
|
"learning_rate": 4.8158233517612096e-05, |
|
"loss": 0.0102, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 0.6733298264071541, |
|
"grad_norm": 1.2987735271453857, |
|
"learning_rate": 4.815276772526186e-05, |
|
"loss": 0.013, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.674031211642995, |
|
"grad_norm": 1.1722874641418457, |
|
"learning_rate": 4.8147294145646207e-05, |
|
"loss": 0.0119, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 0.6747325968788357, |
|
"grad_norm": 30.018028259277344, |
|
"learning_rate": 4.8141812780606137e-05, |
|
"loss": 0.606, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 0.6754339821146765, |
|
"grad_norm": 1.0533334016799927, |
|
"learning_rate": 4.813632363198527e-05, |
|
"loss": 0.0105, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.6761353673505173, |
|
"grad_norm": 52.103050231933594, |
|
"learning_rate": 4.813082670162986e-05, |
|
"loss": 1.5312, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 0.676836752586358, |
|
"grad_norm": 36.39682388305664, |
|
"learning_rate": 4.812532199138873e-05, |
|
"loss": 0.8569, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.6775381378221988, |
|
"grad_norm": 84.3869400024414, |
|
"learning_rate": 4.8119809503113363e-05, |
|
"loss": 2.1346, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 0.6782395230580396, |
|
"grad_norm": 0.8813936710357666, |
|
"learning_rate": 4.811428923865785e-05, |
|
"loss": 0.0085, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 0.6789409082938804, |
|
"grad_norm": 1.1427128314971924, |
|
"learning_rate": 4.810876119987888e-05, |
|
"loss": 0.0113, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 0.6796422935297212, |
|
"grad_norm": 30.400836944580078, |
|
"learning_rate": 4.810322538863578e-05, |
|
"loss": 0.6341, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 0.680343678765562, |
|
"grad_norm": 77.0494155883789, |
|
"learning_rate": 4.8097681806790466e-05, |
|
"loss": 1.7974, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.6810450640014027, |
|
"grad_norm": 29.669004440307617, |
|
"learning_rate": 4.8092130456207495e-05, |
|
"loss": 0.5951, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 0.6817464492372436, |
|
"grad_norm": 2.0304183959960938, |
|
"learning_rate": 4.808657133875402e-05, |
|
"loss": 0.0214, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 0.6824478344730843, |
|
"grad_norm": 22.323291778564453, |
|
"learning_rate": 4.808100445629981e-05, |
|
"loss": 0.4383, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 0.6831492197089252, |
|
"grad_norm": 11.758098602294922, |
|
"learning_rate": 4.807542981071724e-05, |
|
"loss": 0.245, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 0.6838506049447659, |
|
"grad_norm": 9.657792091369629, |
|
"learning_rate": 4.80698474038813e-05, |
|
"loss": 0.2276, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.6845519901806068, |
|
"grad_norm": 0.9010874629020691, |
|
"learning_rate": 4.806425723766961e-05, |
|
"loss": 0.1934, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 0.6852533754164475, |
|
"grad_norm": 13.536153793334961, |
|
"learning_rate": 4.805865931396236e-05, |
|
"loss": 0.2465, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 0.6859547606522882, |
|
"grad_norm": 12.555607795715332, |
|
"learning_rate": 4.8053053634642384e-05, |
|
"loss": 0.2135, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 0.6866561458881291, |
|
"grad_norm": 15.943707466125488, |
|
"learning_rate": 4.804744020159511e-05, |
|
"loss": 0.3215, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 0.6873575311239698, |
|
"grad_norm": 9.25063419342041, |
|
"learning_rate": 4.804181901670856e-05, |
|
"loss": 0.2798, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.6880589163598106, |
|
"grad_norm": 8.65345573425293, |
|
"learning_rate": 4.80361900818734e-05, |
|
"loss": 0.1354, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 0.6887603015956514, |
|
"grad_norm": 10.924114227294922, |
|
"learning_rate": 4.803055339898288e-05, |
|
"loss": 0.1829, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 0.6894616868314922, |
|
"grad_norm": 7.537076950073242, |
|
"learning_rate": 4.8024908969932844e-05, |
|
"loss": 0.1064, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 0.690163072067333, |
|
"grad_norm": 43.258338928222656, |
|
"learning_rate": 4.8019256796621764e-05, |
|
"loss": 0.957, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 0.6908644573031738, |
|
"grad_norm": 26.308055877685547, |
|
"learning_rate": 4.801359688095071e-05, |
|
"loss": 0.541, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.6915658425390145, |
|
"grad_norm": 1.8512808084487915, |
|
"learning_rate": 4.800792922482333e-05, |
|
"loss": 0.0192, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 0.6922672277748554, |
|
"grad_norm": 1.0505925416946411, |
|
"learning_rate": 4.800225383014594e-05, |
|
"loss": 0.0104, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 0.6929686130106961, |
|
"grad_norm": 40.21651840209961, |
|
"learning_rate": 4.799657069882738e-05, |
|
"loss": 1.0056, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 0.6936699982465369, |
|
"grad_norm": 0.17625601589679718, |
|
"learning_rate": 4.7990879832779156e-05, |
|
"loss": 0.0015, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 0.6943713834823777, |
|
"grad_norm": 0.15104326605796814, |
|
"learning_rate": 4.7985181233915335e-05, |
|
"loss": 0.0013, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.6950727687182184, |
|
"grad_norm": 0.09187794476747513, |
|
"learning_rate": 4.79794749041526e-05, |
|
"loss": 0.0008, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 0.6957741539540593, |
|
"grad_norm": 48.85366439819336, |
|
"learning_rate": 4.797376084541024e-05, |
|
"loss": 1.9766, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 0.6964755391899, |
|
"grad_norm": 0.04326195642352104, |
|
"learning_rate": 4.796803905961013e-05, |
|
"loss": 0.0004, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 0.6971769244257409, |
|
"grad_norm": 0.04548042640089989, |
|
"learning_rate": 4.796230954867676e-05, |
|
"loss": 0.0004, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 0.6978783096615816, |
|
"grad_norm": 48.174827575683594, |
|
"learning_rate": 4.79565723145372e-05, |
|
"loss": 1.9204, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.6985796948974224, |
|
"grad_norm": 0.04808010160923004, |
|
"learning_rate": 4.7950827359121135e-05, |
|
"loss": 0.0004, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 0.6992810801332632, |
|
"grad_norm": 48.49928665161133, |
|
"learning_rate": 4.794507468436084e-05, |
|
"loss": 1.8132, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 0.699982465369104, |
|
"grad_norm": 47.405731201171875, |
|
"learning_rate": 4.793931429219117e-05, |
|
"loss": 1.6124, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 0.7006838506049448, |
|
"grad_norm": 0.2506006062030792, |
|
"learning_rate": 4.793354618454961e-05, |
|
"loss": 0.0022, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 0.7013852358407856, |
|
"grad_norm": 90.9857177734375, |
|
"learning_rate": 4.792777036337621e-05, |
|
"loss": 2.7266, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.7020866210766263, |
|
"grad_norm": 0.9927763938903809, |
|
"learning_rate": 4.792198683061363e-05, |
|
"loss": 0.0099, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 0.7027880063124671, |
|
"grad_norm": 1.5031591653823853, |
|
"learning_rate": 4.791619558820712e-05, |
|
"loss": 0.0152, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 0.7034893915483079, |
|
"grad_norm": 29.284543991088867, |
|
"learning_rate": 4.791039663810452e-05, |
|
"loss": 0.5838, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 0.7041907767841487, |
|
"grad_norm": 1.6648647785186768, |
|
"learning_rate": 4.790458998225627e-05, |
|
"loss": 0.0173, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 0.7048921620199895, |
|
"grad_norm": 2.1368696689605713, |
|
"learning_rate": 4.7898775622615384e-05, |
|
"loss": 0.023, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.7055935472558302, |
|
"grad_norm": 1.5531522035598755, |
|
"learning_rate": 4.789295356113749e-05, |
|
"loss": 0.0161, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 0.7062949324916711, |
|
"grad_norm": 1.0346095561981201, |
|
"learning_rate": 4.7887123799780795e-05, |
|
"loss": 0.0096, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 0.7069963177275118, |
|
"grad_norm": 0.5431362986564636, |
|
"learning_rate": 4.788128634050609e-05, |
|
"loss": 0.005, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 0.7076977029633527, |
|
"grad_norm": 0.19359983503818512, |
|
"learning_rate": 4.787544118527677e-05, |
|
"loss": 0.0017, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 0.7083990881991934, |
|
"grad_norm": 49.78437042236328, |
|
"learning_rate": 4.786958833605882e-05, |
|
"loss": 1.7876, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.7091004734350342, |
|
"grad_norm": 57.95826721191406, |
|
"learning_rate": 4.786372779482078e-05, |
|
"loss": 2.9286, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 0.709801858670875, |
|
"grad_norm": 48.60955047607422, |
|
"learning_rate": 4.7857859563533814e-05, |
|
"loss": 1.9456, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 0.7105032439067157, |
|
"grad_norm": 48.62706756591797, |
|
"learning_rate": 4.785198364417166e-05, |
|
"loss": 1.8589, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 0.7112046291425566, |
|
"grad_norm": 94.03495788574219, |
|
"learning_rate": 4.784610003871063e-05, |
|
"loss": 3.2697, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 0.7119060143783973, |
|
"grad_norm": 44.21165084838867, |
|
"learning_rate": 4.784020874912965e-05, |
|
"loss": 1.3436, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.7126073996142381, |
|
"grad_norm": 0.36148276925086975, |
|
"learning_rate": 4.7834309777410195e-05, |
|
"loss": 0.0033, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 0.7133087848500789, |
|
"grad_norm": 74.5863265991211, |
|
"learning_rate": 4.7828403125536354e-05, |
|
"loss": 1.8025, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 0.7140101700859197, |
|
"grad_norm": 43.957252502441406, |
|
"learning_rate": 4.782248879549478e-05, |
|
"loss": 1.1635, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 0.7147115553217604, |
|
"grad_norm": 69.43833923339844, |
|
"learning_rate": 4.781656678927471e-05, |
|
"loss": 1.2, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 0.7154129405576013, |
|
"grad_norm": 4.946931838989258, |
|
"learning_rate": 4.7810637108867986e-05, |
|
"loss": 0.0617, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.716114325793442, |
|
"grad_norm": 2.806675434112549, |
|
"learning_rate": 4.7804699756268987e-05, |
|
"loss": 0.1936, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 0.7168157110292829, |
|
"grad_norm": 10.467972755432129, |
|
"learning_rate": 4.7798754733474724e-05, |
|
"loss": 0.17, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 0.7175170962651236, |
|
"grad_norm": 13.816707611083984, |
|
"learning_rate": 4.7792802042484746e-05, |
|
"loss": 0.2481, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 0.7182184815009645, |
|
"grad_norm": 14.67568588256836, |
|
"learning_rate": 4.77868416853012e-05, |
|
"loss": 0.2728, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 0.7189198667368052, |
|
"grad_norm": 3.8597590923309326, |
|
"learning_rate": 4.778087366392881e-05, |
|
"loss": 0.3473, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.7196212519726459, |
|
"grad_norm": 6.207676410675049, |
|
"learning_rate": 4.777489798037489e-05, |
|
"loss": 0.276, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 0.7203226372084868, |
|
"grad_norm": 1.4748679399490356, |
|
"learning_rate": 4.77689146366493e-05, |
|
"loss": 0.2487, |
|
"step": 1027 |
|
}, |
|
{ |
|
"epoch": 0.7210240224443275, |
|
"grad_norm": 3.2220802307128906, |
|
"learning_rate": 4.77629236347645e-05, |
|
"loss": 0.1941, |
|
"step": 1028 |
|
}, |
|
{ |
|
"epoch": 0.7217254076801684, |
|
"grad_norm": 8.897248268127441, |
|
"learning_rate": 4.775692497673552e-05, |
|
"loss": 0.135, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 0.7224267929160091, |
|
"grad_norm": 6.12791633605957, |
|
"learning_rate": 4.7750918664579956e-05, |
|
"loss": 0.0784, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.7231281781518499, |
|
"grad_norm": 20.013917922973633, |
|
"learning_rate": 4.7744904700318006e-05, |
|
"loss": 0.3768, |
|
"step": 1031 |
|
}, |
|
{ |
|
"epoch": 0.7238295633876907, |
|
"grad_norm": 2.475820541381836, |
|
"learning_rate": 4.7738883085972406e-05, |
|
"loss": 0.0274, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 0.7245309486235315, |
|
"grad_norm": 1.4534077644348145, |
|
"learning_rate": 4.773285382356849e-05, |
|
"loss": 0.015, |
|
"step": 1033 |
|
}, |
|
{ |
|
"epoch": 0.7252323338593722, |
|
"grad_norm": 36.22199630737305, |
|
"learning_rate": 4.772681691513417e-05, |
|
"loss": 0.9046, |
|
"step": 1034 |
|
}, |
|
{ |
|
"epoch": 0.7259337190952131, |
|
"grad_norm": 0.2810744643211365, |
|
"learning_rate": 4.772077236269989e-05, |
|
"loss": 0.0025, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.7266351043310538, |
|
"grad_norm": 45.44533920288086, |
|
"learning_rate": 4.771472016829871e-05, |
|
"loss": 1.3375, |
|
"step": 1036 |
|
}, |
|
{ |
|
"epoch": 0.7273364895668946, |
|
"grad_norm": 0.12535035610198975, |
|
"learning_rate": 4.770866033396623e-05, |
|
"loss": 0.0011, |
|
"step": 1037 |
|
}, |
|
{ |
|
"epoch": 0.7280378748027354, |
|
"grad_norm": 45.64175033569336, |
|
"learning_rate": 4.770259286174065e-05, |
|
"loss": 1.6729, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 0.7287392600385761, |
|
"grad_norm": 46.727394104003906, |
|
"learning_rate": 4.7696517753662696e-05, |
|
"loss": 1.7577, |
|
"step": 1039 |
|
}, |
|
{ |
|
"epoch": 0.729440645274417, |
|
"grad_norm": 0.17975449562072754, |
|
"learning_rate": 4.76904350117757e-05, |
|
"loss": 0.0016, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.7301420305102577, |
|
"grad_norm": 85.70226287841797, |
|
"learning_rate": 4.7684344638125546e-05, |
|
"loss": 2.662, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 0.7308434157460986, |
|
"grad_norm": 39.62570571899414, |
|
"learning_rate": 4.767824663476069e-05, |
|
"loss": 1.1086, |
|
"step": 1042 |
|
}, |
|
{ |
|
"epoch": 0.7315448009819393, |
|
"grad_norm": 33.90923309326172, |
|
"learning_rate": 4.767214100373214e-05, |
|
"loss": 0.8021, |
|
"step": 1043 |
|
}, |
|
{ |
|
"epoch": 0.7322461862177801, |
|
"grad_norm": 23.501230239868164, |
|
"learning_rate": 4.76660277470935e-05, |
|
"loss": 0.4899, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 0.7329475714536209, |
|
"grad_norm": 10.125417709350586, |
|
"learning_rate": 4.76599068669009e-05, |
|
"loss": 0.2671, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.7336489566894617, |
|
"grad_norm": 3.0318775177001953, |
|
"learning_rate": 4.7653778365213076e-05, |
|
"loss": 0.2053, |
|
"step": 1046 |
|
}, |
|
{ |
|
"epoch": 0.7343503419253025, |
|
"grad_norm": 7.598801612854004, |
|
"learning_rate": 4.7647642244091276e-05, |
|
"loss": 0.2937, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 0.7350517271611433, |
|
"grad_norm": 11.337316513061523, |
|
"learning_rate": 4.7641498505599356e-05, |
|
"loss": 0.283, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 0.735753112396984, |
|
"grad_norm": 2.909923553466797, |
|
"learning_rate": 4.763534715180373e-05, |
|
"loss": 0.2887, |
|
"step": 1049 |
|
}, |
|
{ |
|
"epoch": 0.7364544976328248, |
|
"grad_norm": 12.51762866973877, |
|
"learning_rate": 4.762918818477334e-05, |
|
"loss": 0.2306, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.7371558828686656, |
|
"grad_norm": 10.550107955932617, |
|
"learning_rate": 4.762302160657971e-05, |
|
"loss": 0.1879, |
|
"step": 1051 |
|
}, |
|
{ |
|
"epoch": 0.7378572681045064, |
|
"grad_norm": 10.876016616821289, |
|
"learning_rate": 4.7616847419296945e-05, |
|
"loss": 0.1877, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 0.7385586533403472, |
|
"grad_norm": 3.0670106410980225, |
|
"learning_rate": 4.761066562500167e-05, |
|
"loss": 0.2327, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 0.7392600385761879, |
|
"grad_norm": 5.0195770263671875, |
|
"learning_rate": 4.7604476225773084e-05, |
|
"loss": 0.0651, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 0.7399614238120288, |
|
"grad_norm": 2.679926872253418, |
|
"learning_rate": 4.759827922369296e-05, |
|
"loss": 0.0317, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.7406628090478695, |
|
"grad_norm": 1.752316951751709, |
|
"learning_rate": 4.7592074620845616e-05, |
|
"loss": 0.0184, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 0.7413641942837104, |
|
"grad_norm": 0.6544418334960938, |
|
"learning_rate": 4.758586241931791e-05, |
|
"loss": 0.0062, |
|
"step": 1057 |
|
}, |
|
{ |
|
"epoch": 0.7420655795195511, |
|
"grad_norm": 0.19106708467006683, |
|
"learning_rate": 4.757964262119928e-05, |
|
"loss": 0.0017, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 0.742766964755392, |
|
"grad_norm": 0.0741598978638649, |
|
"learning_rate": 4.757341522858171e-05, |
|
"loss": 0.0006, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 0.7434683499912327, |
|
"grad_norm": 0.02510019950568676, |
|
"learning_rate": 4.7567180243559736e-05, |
|
"loss": 0.0002, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.7441697352270735, |
|
"grad_norm": 0.00887396652251482, |
|
"learning_rate": 4.7560937668230454e-05, |
|
"loss": 0.0001, |
|
"step": 1061 |
|
}, |
|
{ |
|
"epoch": 0.7448711204629143, |
|
"grad_norm": 0.0027078278362751007, |
|
"learning_rate": 4.75546875046935e-05, |
|
"loss": 0.0, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 0.745572505698755, |
|
"grad_norm": 0.001226524356752634, |
|
"learning_rate": 4.7548429755051085e-05, |
|
"loss": 0.0, |
|
"step": 1063 |
|
}, |
|
{ |
|
"epoch": 0.7462738909345958, |
|
"grad_norm": 0.0005389907164499164, |
|
"learning_rate": 4.754216442140794e-05, |
|
"loss": 0.0, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 0.7469752761704366, |
|
"grad_norm": 55.77132797241211, |
|
"learning_rate": 4.7535891505871375e-05, |
|
"loss": 3.9293, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.7476766614062774, |
|
"grad_norm": 103.8635025024414, |
|
"learning_rate": 4.752961101055124e-05, |
|
"loss": 7.9354, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 0.7483780466421182, |
|
"grad_norm": 103.83565521240234, |
|
"learning_rate": 4.7523322937559946e-05, |
|
"loss": 7.4002, |
|
"step": 1067 |
|
}, |
|
{ |
|
"epoch": 0.749079431877959, |
|
"grad_norm": 0.0008341918000951409, |
|
"learning_rate": 4.751702728901242e-05, |
|
"loss": 0.0, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 0.7497808171137997, |
|
"grad_norm": 48.793819427490234, |
|
"learning_rate": 4.7510724067026156e-05, |
|
"loss": 3.1709, |
|
"step": 1069 |
|
}, |
|
{ |
|
"epoch": 0.7504822023496406, |
|
"grad_norm": 97.3034439086914, |
|
"learning_rate": 4.750441327372122e-05, |
|
"loss": 5.7, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.7511835875854813, |
|
"grad_norm": 0.0208453219383955, |
|
"learning_rate": 4.7498094911220185e-05, |
|
"loss": 0.0002, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 0.7518849728213222, |
|
"grad_norm": 0.05400345101952553, |
|
"learning_rate": 4.749176898164818e-05, |
|
"loss": 0.0004, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 0.7525863580571629, |
|
"grad_norm": 0.0927615538239479, |
|
"learning_rate": 4.7485435487132904e-05, |
|
"loss": 0.0008, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 0.7532877432930036, |
|
"grad_norm": 0.19018051028251648, |
|
"learning_rate": 4.747909442980456e-05, |
|
"loss": 0.0016, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 0.7539891285288445, |
|
"grad_norm": 0.3820917010307312, |
|
"learning_rate": 4.7472745811795936e-05, |
|
"loss": 0.0035, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.7546905137646852, |
|
"grad_norm": 0.6652111411094666, |
|
"learning_rate": 4.746638963524235e-05, |
|
"loss": 0.0063, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 0.7553918990005261, |
|
"grad_norm": 91.62269592285156, |
|
"learning_rate": 4.7460025902281625e-05, |
|
"loss": 2.2848, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 0.7560932842363668, |
|
"grad_norm": 0.6401788592338562, |
|
"learning_rate": 4.7453654615054174e-05, |
|
"loss": 0.006, |
|
"step": 1078 |
|
}, |
|
{ |
|
"epoch": 0.7567946694722076, |
|
"grad_norm": 0.9835159182548523, |
|
"learning_rate": 4.744727577570295e-05, |
|
"loss": 0.0096, |
|
"step": 1079 |
|
}, |
|
{ |
|
"epoch": 0.7574960547080484, |
|
"grad_norm": 0.665366530418396, |
|
"learning_rate": 4.74408893863734e-05, |
|
"loss": 0.0064, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.7581974399438892, |
|
"grad_norm": 0.3792603313922882, |
|
"learning_rate": 4.743449544921356e-05, |
|
"loss": 0.0034, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 0.75889882517973, |
|
"grad_norm": 0.18616673350334167, |
|
"learning_rate": 4.742809396637397e-05, |
|
"loss": 0.0016, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 0.7596002104155708, |
|
"grad_norm": 0.12753531336784363, |
|
"learning_rate": 4.7421684940007725e-05, |
|
"loss": 0.0011, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 0.7603015956514115, |
|
"grad_norm": 92.24988555908203, |
|
"learning_rate": 4.741526837227047e-05, |
|
"loss": 3.521, |
|
"step": 1084 |
|
}, |
|
{ |
|
"epoch": 0.7610029808872524, |
|
"grad_norm": 48.01845169067383, |
|
"learning_rate": 4.740884426532035e-05, |
|
"loss": 1.9461, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.7617043661230931, |
|
"grad_norm": 0.07205016911029816, |
|
"learning_rate": 4.740241262131808e-05, |
|
"loss": 0.0006, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 0.7624057513589338, |
|
"grad_norm": 0.07788128405809402, |
|
"learning_rate": 4.7395973442426895e-05, |
|
"loss": 0.0007, |
|
"step": 1087 |
|
}, |
|
{ |
|
"epoch": 0.7631071365947747, |
|
"grad_norm": 94.46582794189453, |
|
"learning_rate": 4.738952673081256e-05, |
|
"loss": 3.4385, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 0.7638085218306154, |
|
"grad_norm": 0.17950232326984406, |
|
"learning_rate": 4.738307248864339e-05, |
|
"loss": 0.0016, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 0.7645099070664563, |
|
"grad_norm": 93.51290130615234, |
|
"learning_rate": 4.737661071809021e-05, |
|
"loss": 3.3519, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.765211292302297, |
|
"grad_norm": 92.91374206542969, |
|
"learning_rate": 4.73701414213264e-05, |
|
"loss": 2.7943, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 0.7659126775381379, |
|
"grad_norm": 0.5978343486785889, |
|
"learning_rate": 4.736366460052786e-05, |
|
"loss": 0.0056, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 0.7666140627739786, |
|
"grad_norm": 30.616474151611328, |
|
"learning_rate": 4.7357180257873014e-05, |
|
"loss": 0.6588, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 0.7673154480098194, |
|
"grad_norm": 1.5538511276245117, |
|
"learning_rate": 4.7350688395542825e-05, |
|
"loss": 0.0162, |
|
"step": 1094 |
|
}, |
|
{ |
|
"epoch": 0.7680168332456602, |
|
"grad_norm": 75.00251007080078, |
|
"learning_rate": 4.734418901572079e-05, |
|
"loss": 1.4034, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.768718218481501, |
|
"grad_norm": 3.7650375366210938, |
|
"learning_rate": 4.733768212059293e-05, |
|
"loss": 0.0454, |
|
"step": 1096 |
|
}, |
|
{ |
|
"epoch": 0.7694196037173417, |
|
"grad_norm": 5.952811241149902, |
|
"learning_rate": 4.733116771234778e-05, |
|
"loss": 0.0766, |
|
"step": 1097 |
|
}, |
|
{ |
|
"epoch": 0.7701209889531825, |
|
"grad_norm": 15.600150108337402, |
|
"learning_rate": 4.7324645793176425e-05, |
|
"loss": 0.3215, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 0.7708223741890233, |
|
"grad_norm": 9.213759422302246, |
|
"learning_rate": 4.731811636527246e-05, |
|
"loss": 0.2548, |
|
"step": 1099 |
|
}, |
|
{ |
|
"epoch": 0.7715237594248641, |
|
"grad_norm": 7.49055814743042, |
|
"learning_rate": 4.731157943083202e-05, |
|
"loss": 0.1, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.7722251446607049, |
|
"grad_norm": 7.580368518829346, |
|
"learning_rate": 4.730503499205373e-05, |
|
"loss": 0.1067, |
|
"step": 1101 |
|
}, |
|
{ |
|
"epoch": 0.7729265298965456, |
|
"grad_norm": 6.691344261169434, |
|
"learning_rate": 4.7298483051138796e-05, |
|
"loss": 0.0911, |
|
"step": 1102 |
|
}, |
|
{ |
|
"epoch": 0.7736279151323865, |
|
"grad_norm": 39.10125732421875, |
|
"learning_rate": 4.72919236102909e-05, |
|
"loss": 0.7558, |
|
"step": 1103 |
|
}, |
|
{ |
|
"epoch": 0.7743293003682272, |
|
"grad_norm": 4.285506725311279, |
|
"learning_rate": 4.728535667171626e-05, |
|
"loss": 0.051, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 0.7750306856040681, |
|
"grad_norm": 2.88086199760437, |
|
"learning_rate": 4.727878223762363e-05, |
|
"loss": 0.0324, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.7757320708399088, |
|
"grad_norm": 1.6831408739089966, |
|
"learning_rate": 4.7272200310224256e-05, |
|
"loss": 0.0172, |
|
"step": 1106 |
|
}, |
|
{ |
|
"epoch": 0.7764334560757497, |
|
"grad_norm": 37.848419189453125, |
|
"learning_rate": 4.726561089173194e-05, |
|
"loss": 0.934, |
|
"step": 1107 |
|
}, |
|
{ |
|
"epoch": 0.7771348413115904, |
|
"grad_norm": 38.1249885559082, |
|
"learning_rate": 4.725901398436297e-05, |
|
"loss": 0.9811, |
|
"step": 1108 |
|
}, |
|
{ |
|
"epoch": 0.7778362265474312, |
|
"grad_norm": 51.93446731567383, |
|
"learning_rate": 4.725240959033618e-05, |
|
"loss": 1.5657, |
|
"step": 1109 |
|
}, |
|
{ |
|
"epoch": 0.778537611783272, |
|
"grad_norm": 0.33258679509162903, |
|
"learning_rate": 4.7245797711872895e-05, |
|
"loss": 0.003, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.7792389970191127, |
|
"grad_norm": 61.720924377441406, |
|
"learning_rate": 4.723917835119699e-05, |
|
"loss": 1.6662, |
|
"step": 1111 |
|
}, |
|
{ |
|
"epoch": 0.7799403822549535, |
|
"grad_norm": 37.942874908447266, |
|
"learning_rate": 4.723255151053483e-05, |
|
"loss": 1.0046, |
|
"step": 1112 |
|
}, |
|
{ |
|
"epoch": 0.7806417674907943, |
|
"grad_norm": 0.8135581016540527, |
|
"learning_rate": 4.7225917192115297e-05, |
|
"loss": 0.0076, |
|
"step": 1113 |
|
}, |
|
{ |
|
"epoch": 0.7813431527266351, |
|
"grad_norm": 36.66408920288086, |
|
"learning_rate": 4.721927539816981e-05, |
|
"loss": 0.9099, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.7820445379624759, |
|
"grad_norm": 1.1302094459533691, |
|
"learning_rate": 4.721262613093228e-05, |
|
"loss": 0.0115, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.7827459231983167, |
|
"grad_norm": 1.4476618766784668, |
|
"learning_rate": 4.720596939263914e-05, |
|
"loss": 0.0146, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 0.7834473084341574, |
|
"grad_norm": 1.0591875314712524, |
|
"learning_rate": 4.719930518552934e-05, |
|
"loss": 0.0106, |
|
"step": 1117 |
|
}, |
|
{ |
|
"epoch": 0.7841486936699983, |
|
"grad_norm": 0.7351242899894714, |
|
"learning_rate": 4.7192633511844334e-05, |
|
"loss": 0.0069, |
|
"step": 1118 |
|
}, |
|
{ |
|
"epoch": 0.784850078905839, |
|
"grad_norm": 0.43580272793769836, |
|
"learning_rate": 4.71859543738281e-05, |
|
"loss": 0.004, |
|
"step": 1119 |
|
}, |
|
{ |
|
"epoch": 0.7855514641416799, |
|
"grad_norm": 0.256906658411026, |
|
"learning_rate": 4.71792677737271e-05, |
|
"loss": 0.0022, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.7862528493775206, |
|
"grad_norm": 0.13365890085697174, |
|
"learning_rate": 4.717257371379034e-05, |
|
"loss": 0.0012, |
|
"step": 1121 |
|
}, |
|
{ |
|
"epoch": 0.7869542346133613, |
|
"grad_norm": 0.07990420609712601, |
|
"learning_rate": 4.71658721962693e-05, |
|
"loss": 0.0007, |
|
"step": 1122 |
|
}, |
|
{ |
|
"epoch": 0.7876556198492022, |
|
"grad_norm": 0.0479491651058197, |
|
"learning_rate": 4.715916322341801e-05, |
|
"loss": 0.0004, |
|
"step": 1123 |
|
}, |
|
{ |
|
"epoch": 0.7883570050850429, |
|
"grad_norm": 47.28627014160156, |
|
"learning_rate": 4.7152446797492963e-05, |
|
"loss": 2.0982, |
|
"step": 1124 |
|
}, |
|
{ |
|
"epoch": 0.7890583903208838, |
|
"grad_norm": 96.63157653808594, |
|
"learning_rate": 4.71457229207532e-05, |
|
"loss": 4.5257, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.7897597755567245, |
|
"grad_norm": 47.81488800048828, |
|
"learning_rate": 4.713899159546024e-05, |
|
"loss": 2.3268, |
|
"step": 1126 |
|
}, |
|
{ |
|
"epoch": 0.7904611607925653, |
|
"grad_norm": 47.34000015258789, |
|
"learning_rate": 4.713225282387811e-05, |
|
"loss": 2.1484, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 0.7911625460284061, |
|
"grad_norm": 46.76436996459961, |
|
"learning_rate": 4.7125506608273344e-05, |
|
"loss": 1.8822, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 0.7918639312642469, |
|
"grad_norm": 0.12850354611873627, |
|
"learning_rate": 4.711875295091499e-05, |
|
"loss": 0.0011, |
|
"step": 1129 |
|
}, |
|
{ |
|
"epoch": 0.7925653165000877, |
|
"grad_norm": 126.00606536865234, |
|
"learning_rate": 4.71119918540746e-05, |
|
"loss": 3.7093, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.7932667017359285, |
|
"grad_norm": 0.4526232182979584, |
|
"learning_rate": 4.7105223320026205e-05, |
|
"loss": 0.0042, |
|
"step": 1131 |
|
}, |
|
{ |
|
"epoch": 0.7939680869717692, |
|
"grad_norm": 33.449928283691406, |
|
"learning_rate": 4.709844735104636e-05, |
|
"loss": 0.7578, |
|
"step": 1132 |
|
}, |
|
{ |
|
"epoch": 0.7946694722076101, |
|
"grad_norm": 1.6459629535675049, |
|
"learning_rate": 4.7091663949414105e-05, |
|
"loss": 0.0174, |
|
"step": 1133 |
|
}, |
|
{ |
|
"epoch": 0.7953708574434508, |
|
"grad_norm": 34.26624298095703, |
|
"learning_rate": 4.7084873117411e-05, |
|
"loss": 0.7867, |
|
"step": 1134 |
|
}, |
|
{ |
|
"epoch": 0.7960722426792916, |
|
"grad_norm": 3.238154888153076, |
|
"learning_rate": 4.7078074857321085e-05, |
|
"loss": 0.0382, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.7967736279151324, |
|
"grad_norm": 3.573115348815918, |
|
"learning_rate": 4.70712691714309e-05, |
|
"loss": 0.0431, |
|
"step": 1136 |
|
}, |
|
{ |
|
"epoch": 0.7974750131509731, |
|
"grad_norm": 4.002458095550537, |
|
"learning_rate": 4.7064456062029506e-05, |
|
"loss": 0.0488, |
|
"step": 1137 |
|
}, |
|
{ |
|
"epoch": 0.798176398386814, |
|
"grad_norm": 3.4627814292907715, |
|
"learning_rate": 4.7057635531408424e-05, |
|
"loss": 0.0394, |
|
"step": 1138 |
|
}, |
|
{ |
|
"epoch": 0.7988777836226547, |
|
"grad_norm": 18.19850730895996, |
|
"learning_rate": 4.70508075818617e-05, |
|
"loss": 0.3679, |
|
"step": 1139 |
|
}, |
|
{ |
|
"epoch": 0.7995791688584956, |
|
"grad_norm": 2.7357804775238037, |
|
"learning_rate": 4.704397221568586e-05, |
|
"loss": 0.0312, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.8002805540943363, |
|
"grad_norm": 51.47498321533203, |
|
"learning_rate": 4.703712943517993e-05, |
|
"loss": 1.0259, |
|
"step": 1141 |
|
}, |
|
{ |
|
"epoch": 0.8009819393301771, |
|
"grad_norm": 1.7935067415237427, |
|
"learning_rate": 4.7030279242645434e-05, |
|
"loss": 0.0192, |
|
"step": 1142 |
|
}, |
|
{ |
|
"epoch": 0.8016833245660179, |
|
"grad_norm": 28.79102325439453, |
|
"learning_rate": 4.702342164038639e-05, |
|
"loss": 0.6393, |
|
"step": 1143 |
|
}, |
|
{ |
|
"epoch": 0.8023847098018587, |
|
"grad_norm": 1.230648398399353, |
|
"learning_rate": 4.7016556630709284e-05, |
|
"loss": 0.0126, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 0.8030860950376995, |
|
"grad_norm": 0.9287470579147339, |
|
"learning_rate": 4.700968421592313e-05, |
|
"loss": 0.0094, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.8037874802735402, |
|
"grad_norm": 62.944549560546875, |
|
"learning_rate": 4.700280439833941e-05, |
|
"loss": 1.3975, |
|
"step": 1146 |
|
}, |
|
{ |
|
"epoch": 0.804488865509381, |
|
"grad_norm": 44.87819290161133, |
|
"learning_rate": 4.69959171802721e-05, |
|
"loss": 1.0857, |
|
"step": 1147 |
|
}, |
|
{ |
|
"epoch": 0.8051902507452218, |
|
"grad_norm": 0.5283865928649902, |
|
"learning_rate": 4.6989022564037654e-05, |
|
"loss": 0.0049, |
|
"step": 1148 |
|
}, |
|
{ |
|
"epoch": 0.8058916359810626, |
|
"grad_norm": 0.5992218852043152, |
|
"learning_rate": 4.698212055195505e-05, |
|
"loss": 0.0058, |
|
"step": 1149 |
|
}, |
|
{ |
|
"epoch": 0.8065930212169033, |
|
"grad_norm": 0.5411283373832703, |
|
"learning_rate": 4.697521114634571e-05, |
|
"loss": 0.0052, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.8072944064527442, |
|
"grad_norm": 75.1710433959961, |
|
"learning_rate": 4.696829434953357e-05, |
|
"loss": 1.9684, |
|
"step": 1151 |
|
}, |
|
{ |
|
"epoch": 0.8079957916885849, |
|
"grad_norm": 0.3737725615501404, |
|
"learning_rate": 4.6961370163845044e-05, |
|
"loss": 0.0035, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 0.8086971769244258, |
|
"grad_norm": 48.369232177734375, |
|
"learning_rate": 4.6954438591609027e-05, |
|
"loss": 1.2948, |
|
"step": 1153 |
|
}, |
|
{ |
|
"epoch": 0.8093985621602665, |
|
"grad_norm": 0.4170537292957306, |
|
"learning_rate": 4.694749963515691e-05, |
|
"loss": 0.0036, |
|
"step": 1154 |
|
}, |
|
{ |
|
"epoch": 0.8100999473961074, |
|
"grad_norm": 0.3476961553096771, |
|
"learning_rate": 4.6940553296822565e-05, |
|
"loss": 0.0031, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.8108013326319481, |
|
"grad_norm": 40.558204650878906, |
|
"learning_rate": 4.693359957894232e-05, |
|
"loss": 1.199, |
|
"step": 1156 |
|
}, |
|
{ |
|
"epoch": 0.8115027178677889, |
|
"grad_norm": 78.08576202392578, |
|
"learning_rate": 4.692663848385503e-05, |
|
"loss": 2.1697, |
|
"step": 1157 |
|
}, |
|
{ |
|
"epoch": 0.8122041031036297, |
|
"grad_norm": 0.5332819819450378, |
|
"learning_rate": 4.6919670013902015e-05, |
|
"loss": 0.0051, |
|
"step": 1158 |
|
}, |
|
{ |
|
"epoch": 0.8129054883394704, |
|
"grad_norm": 0.6460719108581543, |
|
"learning_rate": 4.691269417142704e-05, |
|
"loss": 0.0063, |
|
"step": 1159 |
|
}, |
|
{ |
|
"epoch": 0.8136068735753113, |
|
"grad_norm": 0.6999556422233582, |
|
"learning_rate": 4.69057109587764e-05, |
|
"loss": 0.0068, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.814308258811152, |
|
"grad_norm": 0.5912299156188965, |
|
"learning_rate": 4.6898720378298846e-05, |
|
"loss": 0.0057, |
|
"step": 1161 |
|
}, |
|
{ |
|
"epoch": 0.8150096440469928, |
|
"grad_norm": 0.4620411992073059, |
|
"learning_rate": 4.6891722432345597e-05, |
|
"loss": 0.0044, |
|
"step": 1162 |
|
}, |
|
{ |
|
"epoch": 0.8157110292828336, |
|
"grad_norm": 0.3710283935070038, |
|
"learning_rate": 4.6884717123270374e-05, |
|
"loss": 0.0034, |
|
"step": 1163 |
|
}, |
|
{ |
|
"epoch": 0.8164124145186744, |
|
"grad_norm": 0.22289921343326569, |
|
"learning_rate": 4.687770445342936e-05, |
|
"loss": 0.002, |
|
"step": 1164 |
|
}, |
|
{ |
|
"epoch": 0.8171137997545151, |
|
"grad_norm": 89.52765655517578, |
|
"learning_rate": 4.687068442518121e-05, |
|
"loss": 2.8479, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.817815184990356, |
|
"grad_norm": 0.15177088975906372, |
|
"learning_rate": 4.6863657040887065e-05, |
|
"loss": 0.0013, |
|
"step": 1166 |
|
}, |
|
{ |
|
"epoch": 0.8185165702261967, |
|
"grad_norm": 0.13902179896831512, |
|
"learning_rate": 4.685662230291054e-05, |
|
"loss": 0.0012, |
|
"step": 1167 |
|
}, |
|
{ |
|
"epoch": 0.8192179554620376, |
|
"grad_norm": 0.10139818489551544, |
|
"learning_rate": 4.68495802136177e-05, |
|
"loss": 0.0009, |
|
"step": 1168 |
|
}, |
|
{ |
|
"epoch": 0.8199193406978783, |
|
"grad_norm": 0.08662685006856918, |
|
"learning_rate": 4.684253077537712e-05, |
|
"loss": 0.0007, |
|
"step": 1169 |
|
}, |
|
{ |
|
"epoch": 0.820620725933719, |
|
"grad_norm": 0.07555234432220459, |
|
"learning_rate": 4.683547399055982e-05, |
|
"loss": 0.0006, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.8213221111695599, |
|
"grad_norm": 0.06445518881082535, |
|
"learning_rate": 4.682840986153929e-05, |
|
"loss": 0.0006, |
|
"step": 1171 |
|
}, |
|
{ |
|
"epoch": 0.8220234964054006, |
|
"grad_norm": 46.72000503540039, |
|
"learning_rate": 4.68213383906915e-05, |
|
"loss": 2.1281, |
|
"step": 1172 |
|
}, |
|
{ |
|
"epoch": 0.8227248816412415, |
|
"grad_norm": 46.803985595703125, |
|
"learning_rate": 4.6814259580394915e-05, |
|
"loss": 1.9621, |
|
"step": 1173 |
|
}, |
|
{ |
|
"epoch": 0.8234262668770822, |
|
"grad_norm": 0.07187401503324509, |
|
"learning_rate": 4.68071734330304e-05, |
|
"loss": 0.0006, |
|
"step": 1174 |
|
}, |
|
{ |
|
"epoch": 0.824127652112923, |
|
"grad_norm": 0.09642617404460907, |
|
"learning_rate": 4.680007995098136e-05, |
|
"loss": 0.0008, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.8248290373487638, |
|
"grad_norm": 0.11466558277606964, |
|
"learning_rate": 4.679297913663363e-05, |
|
"loss": 0.001, |
|
"step": 1176 |
|
}, |
|
{ |
|
"epoch": 0.8255304225846046, |
|
"grad_norm": 0.11334109306335449, |
|
"learning_rate": 4.67858709923755e-05, |
|
"loss": 0.001, |
|
"step": 1177 |
|
}, |
|
{ |
|
"epoch": 0.8262318078204454, |
|
"grad_norm": 0.1963043361902237, |
|
"learning_rate": 4.677875552059777e-05, |
|
"loss": 0.0017, |
|
"step": 1178 |
|
}, |
|
{ |
|
"epoch": 0.8269331930562862, |
|
"grad_norm": 0.17632529139518738, |
|
"learning_rate": 4.677163272369365e-05, |
|
"loss": 0.0016, |
|
"step": 1179 |
|
}, |
|
{ |
|
"epoch": 0.8276345782921269, |
|
"grad_norm": 124.80657958984375, |
|
"learning_rate": 4.6764502604058864e-05, |
|
"loss": 3.9859, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.8283359635279678, |
|
"grad_norm": 0.19632282853126526, |
|
"learning_rate": 4.675736516409156e-05, |
|
"loss": 0.0018, |
|
"step": 1181 |
|
}, |
|
{ |
|
"epoch": 0.8290373487638085, |
|
"grad_norm": 43.057350158691406, |
|
"learning_rate": 4.675022040619238e-05, |
|
"loss": 1.2696, |
|
"step": 1182 |
|
}, |
|
{ |
|
"epoch": 0.8297387339996493, |
|
"grad_norm": 39.46339416503906, |
|
"learning_rate": 4.674306833276439e-05, |
|
"loss": 1.093, |
|
"step": 1183 |
|
}, |
|
{ |
|
"epoch": 0.8304401192354901, |
|
"grad_norm": 0.653581976890564, |
|
"learning_rate": 4.6735908946213156e-05, |
|
"loss": 0.0065, |
|
"step": 1184 |
|
}, |
|
{ |
|
"epoch": 0.8311415044713308, |
|
"grad_norm": 0.8602274656295776, |
|
"learning_rate": 4.6728742248946677e-05, |
|
"loss": 0.0088, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.8318428897071717, |
|
"grad_norm": 0.9501532912254333, |
|
"learning_rate": 4.6721568243375425e-05, |
|
"loss": 0.0095, |
|
"step": 1186 |
|
}, |
|
{ |
|
"epoch": 0.8325442749430124, |
|
"grad_norm": 0.8702390193939209, |
|
"learning_rate": 4.671438693191232e-05, |
|
"loss": 0.0088, |
|
"step": 1187 |
|
}, |
|
{ |
|
"epoch": 0.8332456601788533, |
|
"grad_norm": 0.6245111227035522, |
|
"learning_rate": 4.6707198316972746e-05, |
|
"loss": 0.0061, |
|
"step": 1188 |
|
}, |
|
{ |
|
"epoch": 0.833947045414694, |
|
"grad_norm": 0.4615590572357178, |
|
"learning_rate": 4.6700002400974544e-05, |
|
"loss": 0.0045, |
|
"step": 1189 |
|
}, |
|
{ |
|
"epoch": 0.8346484306505348, |
|
"grad_norm": 0.3482006788253784, |
|
"learning_rate": 4.6692799186338e-05, |
|
"loss": 0.0033, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.8353498158863756, |
|
"grad_norm": 43.396461486816406, |
|
"learning_rate": 4.668558867548587e-05, |
|
"loss": 1.3748, |
|
"step": 1191 |
|
}, |
|
{ |
|
"epoch": 0.8360512011222164, |
|
"grad_norm": 49.66410827636719, |
|
"learning_rate": 4.6678370870843357e-05, |
|
"loss": 1.7113, |
|
"step": 1192 |
|
}, |
|
{ |
|
"epoch": 0.8367525863580572, |
|
"grad_norm": 0.1562856137752533, |
|
"learning_rate": 4.667114577483812e-05, |
|
"loss": 0.0014, |
|
"step": 1193 |
|
}, |
|
{ |
|
"epoch": 0.8374539715938979, |
|
"grad_norm": 47.60147476196289, |
|
"learning_rate": 4.666391338990026e-05, |
|
"loss": 1.8113, |
|
"step": 1194 |
|
}, |
|
{ |
|
"epoch": 0.8381553568297387, |
|
"grad_norm": 0.19547221064567566, |
|
"learning_rate": 4.665667371846234e-05, |
|
"loss": 0.0018, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.8388567420655795, |
|
"grad_norm": 0.23496605455875397, |
|
"learning_rate": 4.664942676295937e-05, |
|
"loss": 0.0022, |
|
"step": 1196 |
|
}, |
|
{ |
|
"epoch": 0.8395581273014203, |
|
"grad_norm": 0.19859689474105835, |
|
"learning_rate": 4.6642172525828806e-05, |
|
"loss": 0.0018, |
|
"step": 1197 |
|
}, |
|
{ |
|
"epoch": 0.840259512537261, |
|
"grad_norm": 0.21561066806316376, |
|
"learning_rate": 4.6634911009510574e-05, |
|
"loss": 0.002, |
|
"step": 1198 |
|
}, |
|
{ |
|
"epoch": 0.8409608977731019, |
|
"grad_norm": 0.16780033707618713, |
|
"learning_rate": 4.662764221644702e-05, |
|
"loss": 0.0016, |
|
"step": 1199 |
|
}, |
|
{ |
|
"epoch": 0.8416622830089426, |
|
"grad_norm": 0.170645073056221, |
|
"learning_rate": 4.6620366149082954e-05, |
|
"loss": 0.0015, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.8423636682447835, |
|
"grad_norm": 0.16811063885688782, |
|
"learning_rate": 4.661308280986563e-05, |
|
"loss": 0.0015, |
|
"step": 1201 |
|
}, |
|
{ |
|
"epoch": 0.8430650534806242, |
|
"grad_norm": 42.41083908081055, |
|
"learning_rate": 4.6605792201244733e-05, |
|
"loss": 1.4838, |
|
"step": 1202 |
|
}, |
|
{ |
|
"epoch": 0.8437664387164651, |
|
"grad_norm": 43.9367790222168, |
|
"learning_rate": 4.6598494325672426e-05, |
|
"loss": 1.4149, |
|
"step": 1203 |
|
}, |
|
{ |
|
"epoch": 0.8444678239523058, |
|
"grad_norm": 0.1912125051021576, |
|
"learning_rate": 4.659118918560329e-05, |
|
"loss": 0.0018, |
|
"step": 1204 |
|
}, |
|
{ |
|
"epoch": 0.8451692091881466, |
|
"grad_norm": 0.27999013662338257, |
|
"learning_rate": 4.6583876783494343e-05, |
|
"loss": 0.0025, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.8458705944239874, |
|
"grad_norm": 36.06639099121094, |
|
"learning_rate": 4.657655712180507e-05, |
|
"loss": 0.9947, |
|
"step": 1206 |
|
}, |
|
{ |
|
"epoch": 0.8465719796598281, |
|
"grad_norm": 0.4413144588470459, |
|
"learning_rate": 4.6569230202997385e-05, |
|
"loss": 0.0042, |
|
"step": 1207 |
|
}, |
|
{ |
|
"epoch": 0.847273364895669, |
|
"grad_norm": 0.530229926109314, |
|
"learning_rate": 4.656189602953565e-05, |
|
"loss": 0.0052, |
|
"step": 1208 |
|
}, |
|
{ |
|
"epoch": 0.8479747501315097, |
|
"grad_norm": 0.6561914086341858, |
|
"learning_rate": 4.655455460388665e-05, |
|
"loss": 0.0066, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 0.8486761353673505, |
|
"grad_norm": 0.6459203362464905, |
|
"learning_rate": 4.654720592851962e-05, |
|
"loss": 0.0063, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.8493775206031913, |
|
"grad_norm": 37.68455505371094, |
|
"learning_rate": 4.653985000590624e-05, |
|
"loss": 1.0854, |
|
"step": 1211 |
|
}, |
|
{ |
|
"epoch": 0.8500789058390321, |
|
"grad_norm": 0.4941748380661011, |
|
"learning_rate": 4.6532486838520616e-05, |
|
"loss": 0.0049, |
|
"step": 1212 |
|
}, |
|
{ |
|
"epoch": 0.8507802910748729, |
|
"grad_norm": 38.626068115234375, |
|
"learning_rate": 4.6525116428839307e-05, |
|
"loss": 1.0544, |
|
"step": 1213 |
|
}, |
|
{ |
|
"epoch": 0.8514816763107137, |
|
"grad_norm": 0.5088845491409302, |
|
"learning_rate": 4.651773877934129e-05, |
|
"loss": 0.005, |
|
"step": 1214 |
|
}, |
|
{ |
|
"epoch": 0.8521830615465544, |
|
"grad_norm": 0.5165444016456604, |
|
"learning_rate": 4.651035389250798e-05, |
|
"loss": 0.005, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.8528844467823953, |
|
"grad_norm": 37.2906379699707, |
|
"learning_rate": 4.650296177082323e-05, |
|
"loss": 1.0227, |
|
"step": 1216 |
|
}, |
|
{ |
|
"epoch": 0.853585832018236, |
|
"grad_norm": 0.5255476832389832, |
|
"learning_rate": 4.649556241677333e-05, |
|
"loss": 0.0052, |
|
"step": 1217 |
|
}, |
|
{ |
|
"epoch": 0.8542872172540767, |
|
"grad_norm": 37.65691375732422, |
|
"learning_rate": 4.6488155832847006e-05, |
|
"loss": 1.1044, |
|
"step": 1218 |
|
}, |
|
{ |
|
"epoch": 0.8549886024899176, |
|
"grad_norm": 35.80206298828125, |
|
"learning_rate": 4.648074202153539e-05, |
|
"loss": 0.8649, |
|
"step": 1219 |
|
}, |
|
{ |
|
"epoch": 0.8556899877257583, |
|
"grad_norm": 0.8690975308418274, |
|
"learning_rate": 4.647332098533208e-05, |
|
"loss": 0.0089, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.8563913729615992, |
|
"grad_norm": 1.0327463150024414, |
|
"learning_rate": 4.646589272673308e-05, |
|
"loss": 0.0109, |
|
"step": 1221 |
|
}, |
|
{ |
|
"epoch": 0.8570927581974399, |
|
"grad_norm": 32.14468765258789, |
|
"learning_rate": 4.645845724823684e-05, |
|
"loss": 0.7778, |
|
"step": 1222 |
|
}, |
|
{ |
|
"epoch": 0.8577941434332808, |
|
"grad_norm": 1.0274231433868408, |
|
"learning_rate": 4.645101455234422e-05, |
|
"loss": 0.0106, |
|
"step": 1223 |
|
}, |
|
{ |
|
"epoch": 0.8584955286691215, |
|
"grad_norm": 29.724233627319336, |
|
"learning_rate": 4.644356464155852e-05, |
|
"loss": 0.6951, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 0.8591969139049623, |
|
"grad_norm": 1.3864684104919434, |
|
"learning_rate": 4.6436107518385465e-05, |
|
"loss": 0.0151, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.8598982991408031, |
|
"grad_norm": 1.3250905275344849, |
|
"learning_rate": 4.6428643185333196e-05, |
|
"loss": 0.0144, |
|
"step": 1226 |
|
}, |
|
{ |
|
"epoch": 0.8605996843766439, |
|
"grad_norm": 29.683332443237305, |
|
"learning_rate": 4.6421171644912295e-05, |
|
"loss": 0.7007, |
|
"step": 1227 |
|
}, |
|
{ |
|
"epoch": 0.8613010696124846, |
|
"grad_norm": 29.98341178894043, |
|
"learning_rate": 4.6413692899635764e-05, |
|
"loss": 0.6981, |
|
"step": 1228 |
|
}, |
|
{ |
|
"epoch": 0.8620024548483255, |
|
"grad_norm": 30.50311279296875, |
|
"learning_rate": 4.6406206952019014e-05, |
|
"loss": 0.7341, |
|
"step": 1229 |
|
}, |
|
{ |
|
"epoch": 0.8627038400841662, |
|
"grad_norm": 1.5321385860443115, |
|
"learning_rate": 4.6398713804579896e-05, |
|
"loss": 0.0171, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.863405225320007, |
|
"grad_norm": 25.173595428466797, |
|
"learning_rate": 4.6391213459838676e-05, |
|
"loss": 0.5393, |
|
"step": 1231 |
|
}, |
|
{ |
|
"epoch": 0.8641066105558478, |
|
"grad_norm": 1.9056795835494995, |
|
"learning_rate": 4.638370592031804e-05, |
|
"loss": 0.0216, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 0.8648079957916885, |
|
"grad_norm": 54.3871955871582, |
|
"learning_rate": 4.63761911885431e-05, |
|
"loss": 1.1726, |
|
"step": 1233 |
|
}, |
|
{ |
|
"epoch": 0.8655093810275294, |
|
"grad_norm": 2.3975331783294678, |
|
"learning_rate": 4.6368669267041364e-05, |
|
"loss": 0.0284, |
|
"step": 1234 |
|
}, |
|
{ |
|
"epoch": 0.8662107662633701, |
|
"grad_norm": 2.3505735397338867, |
|
"learning_rate": 4.636114015834279e-05, |
|
"loss": 0.027, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.866912151499211, |
|
"grad_norm": 19.353313446044922, |
|
"learning_rate": 4.635360386497973e-05, |
|
"loss": 0.4174, |
|
"step": 1236 |
|
}, |
|
{ |
|
"epoch": 0.8676135367350517, |
|
"grad_norm": 2.5313189029693604, |
|
"learning_rate": 4.6346060389486976e-05, |
|
"loss": 0.0301, |
|
"step": 1237 |
|
}, |
|
{ |
|
"epoch": 0.8683149219708926, |
|
"grad_norm": 2.249114751815796, |
|
"learning_rate": 4.633850973440171e-05, |
|
"loss": 0.0262, |
|
"step": 1238 |
|
}, |
|
{ |
|
"epoch": 0.8690163072067333, |
|
"grad_norm": 1.087750792503357, |
|
"learning_rate": 4.633095190226355e-05, |
|
"loss": 0.0116, |
|
"step": 1239 |
|
}, |
|
{ |
|
"epoch": 0.8697176924425741, |
|
"grad_norm": 0.7214330434799194, |
|
"learning_rate": 4.63233868956145e-05, |
|
"loss": 0.0075, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.8704190776784149, |
|
"grad_norm": 0.5183669924736023, |
|
"learning_rate": 4.6315814716999003e-05, |
|
"loss": 0.0051, |
|
"step": 1241 |
|
}, |
|
{ |
|
"epoch": 0.8711204629142556, |
|
"grad_norm": 0.26819220185279846, |
|
"learning_rate": 4.630823536896392e-05, |
|
"loss": 0.0026, |
|
"step": 1242 |
|
}, |
|
{ |
|
"epoch": 0.8718218481500964, |
|
"grad_norm": 42.74302291870117, |
|
"learning_rate": 4.6300648854058496e-05, |
|
"loss": 1.4745, |
|
"step": 1243 |
|
}, |
|
{ |
|
"epoch": 0.8725232333859372, |
|
"grad_norm": 0.06916595250368118, |
|
"learning_rate": 4.62930551748344e-05, |
|
"loss": 0.0006, |
|
"step": 1244 |
|
}, |
|
{ |
|
"epoch": 0.873224618621778, |
|
"grad_norm": 0.04199717566370964, |
|
"learning_rate": 4.6285454333845714e-05, |
|
"loss": 0.0004, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.8739260038576188, |
|
"grad_norm": 46.651554107666016, |
|
"learning_rate": 4.6277846333648926e-05, |
|
"loss": 2.1377, |
|
"step": 1246 |
|
}, |
|
{ |
|
"epoch": 0.8746273890934596, |
|
"grad_norm": 0.019042061641812325, |
|
"learning_rate": 4.6270231176802936e-05, |
|
"loss": 0.0002, |
|
"step": 1247 |
|
}, |
|
{ |
|
"epoch": 0.8753287743293003, |
|
"grad_norm": 43.70697784423828, |
|
"learning_rate": 4.626260886586903e-05, |
|
"loss": 1.9754, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 0.8760301595651412, |
|
"grad_norm": 0.03968697413802147, |
|
"learning_rate": 4.625497940341094e-05, |
|
"loss": 0.0003, |
|
"step": 1249 |
|
}, |
|
{ |
|
"epoch": 0.8767315448009819, |
|
"grad_norm": 0.023654410615563393, |
|
"learning_rate": 4.624734279199476e-05, |
|
"loss": 0.0002, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.8774329300368228, |
|
"grad_norm": 0.05060436204075813, |
|
"learning_rate": 4.623969903418903e-05, |
|
"loss": 0.0004, |
|
"step": 1251 |
|
}, |
|
{ |
|
"epoch": 0.8781343152726635, |
|
"grad_norm": 0.062442123889923096, |
|
"learning_rate": 4.623204813256465e-05, |
|
"loss": 0.0006, |
|
"step": 1252 |
|
}, |
|
{ |
|
"epoch": 0.8788357005085043, |
|
"grad_norm": 0.05655084550380707, |
|
"learning_rate": 4.622439008969497e-05, |
|
"loss": 0.0005, |
|
"step": 1253 |
|
}, |
|
{ |
|
"epoch": 0.8795370857443451, |
|
"grad_norm": 0.05976495519280434, |
|
"learning_rate": 4.621672490815569e-05, |
|
"loss": 0.0005, |
|
"step": 1254 |
|
}, |
|
{ |
|
"epoch": 0.8802384709801858, |
|
"grad_norm": 0.06660816818475723, |
|
"learning_rate": 4.620905259052496e-05, |
|
"loss": 0.0006, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.8809398562160267, |
|
"grad_norm": 45.397789001464844, |
|
"learning_rate": 4.6201373139383284e-05, |
|
"loss": 1.7976, |
|
"step": 1256 |
|
}, |
|
{ |
|
"epoch": 0.8816412414518674, |
|
"grad_norm": 44.1479606628418, |
|
"learning_rate": 4.619368655731362e-05, |
|
"loss": 1.7646, |
|
"step": 1257 |
|
}, |
|
{ |
|
"epoch": 0.8823426266877082, |
|
"grad_norm": 0.13133244216442108, |
|
"learning_rate": 4.618599284690127e-05, |
|
"loss": 0.0012, |
|
"step": 1258 |
|
}, |
|
{ |
|
"epoch": 0.883044011923549, |
|
"grad_norm": 82.86041259765625, |
|
"learning_rate": 4.6178292010733984e-05, |
|
"loss": 2.727, |
|
"step": 1259 |
|
}, |
|
{ |
|
"epoch": 0.8837453971593898, |
|
"grad_norm": 36.786067962646484, |
|
"learning_rate": 4.6170584051401846e-05, |
|
"loss": 1.0486, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.8844467823952306, |
|
"grad_norm": 43.94260787963867, |
|
"learning_rate": 4.6162868971497396e-05, |
|
"loss": 1.3111, |
|
"step": 1261 |
|
}, |
|
{ |
|
"epoch": 0.8851481676310714, |
|
"grad_norm": 85.85136413574219, |
|
"learning_rate": 4.615514677361554e-05, |
|
"loss": 2.2235, |
|
"step": 1262 |
|
}, |
|
{ |
|
"epoch": 0.8858495528669121, |
|
"grad_norm": 35.138553619384766, |
|
"learning_rate": 4.6147417460353596e-05, |
|
"loss": 0.6498, |
|
"step": 1263 |
|
}, |
|
{ |
|
"epoch": 0.886550938102753, |
|
"grad_norm": 5.97971773147583, |
|
"learning_rate": 4.613968103431124e-05, |
|
"loss": 0.0857, |
|
"step": 1264 |
|
}, |
|
{ |
|
"epoch": 0.8872523233385937, |
|
"grad_norm": 7.489769458770752, |
|
"learning_rate": 4.6131937498090575e-05, |
|
"loss": 0.1154, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.8879537085744345, |
|
"grad_norm": 9.91996955871582, |
|
"learning_rate": 4.612418685429608e-05, |
|
"loss": 0.1685, |
|
"step": 1266 |
|
}, |
|
{ |
|
"epoch": 0.8886550938102753, |
|
"grad_norm": 2.7866861820220947, |
|
"learning_rate": 4.611642910553463e-05, |
|
"loss": 0.2187, |
|
"step": 1267 |
|
}, |
|
{ |
|
"epoch": 0.889356479046116, |
|
"grad_norm": 9.682534217834473, |
|
"learning_rate": 4.610866425441549e-05, |
|
"loss": 0.1669, |
|
"step": 1268 |
|
}, |
|
{ |
|
"epoch": 0.8900578642819569, |
|
"grad_norm": 8.847615242004395, |
|
"learning_rate": 4.6100892303550305e-05, |
|
"loss": 0.1458, |
|
"step": 1269 |
|
}, |
|
{ |
|
"epoch": 0.8907592495177976, |
|
"grad_norm": 6.194153308868408, |
|
"learning_rate": 4.609311325555311e-05, |
|
"loss": 0.0885, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.8914606347536385, |
|
"grad_norm": 11.383880615234375, |
|
"learning_rate": 4.6085327113040346e-05, |
|
"loss": 0.2945, |
|
"step": 1271 |
|
}, |
|
{ |
|
"epoch": 0.8921620199894792, |
|
"grad_norm": 41.971961975097656, |
|
"learning_rate": 4.607753387863082e-05, |
|
"loss": 0.8504, |
|
"step": 1272 |
|
}, |
|
{ |
|
"epoch": 0.89286340522532, |
|
"grad_norm": 15.673029899597168, |
|
"learning_rate": 4.6069733554945725e-05, |
|
"loss": 0.3576, |
|
"step": 1273 |
|
}, |
|
{ |
|
"epoch": 0.8935647904611608, |
|
"grad_norm": 2.994936466217041, |
|
"learning_rate": 4.606192614460865e-05, |
|
"loss": 0.0371, |
|
"step": 1274 |
|
}, |
|
{ |
|
"epoch": 0.8942661756970016, |
|
"grad_norm": 15.890610694885254, |
|
"learning_rate": 4.6054111650245544e-05, |
|
"loss": 0.3475, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.8949675609328424, |
|
"grad_norm": 3.3848869800567627, |
|
"learning_rate": 4.604629007448479e-05, |
|
"loss": 0.0433, |
|
"step": 1276 |
|
}, |
|
{ |
|
"epoch": 0.8956689461686832, |
|
"grad_norm": 26.11835289001465, |
|
"learning_rate": 4.603846141995708e-05, |
|
"loss": 0.5389, |
|
"step": 1277 |
|
}, |
|
{ |
|
"epoch": 0.8963703314045239, |
|
"grad_norm": 3.1392459869384766, |
|
"learning_rate": 4.6030625689295546e-05, |
|
"loss": 0.0383, |
|
"step": 1278 |
|
}, |
|
{ |
|
"epoch": 0.8970717166403647, |
|
"grad_norm": 3.49643611907959, |
|
"learning_rate": 4.6022782885135674e-05, |
|
"loss": 0.0445, |
|
"step": 1279 |
|
}, |
|
{ |
|
"epoch": 0.8977731018762055, |
|
"grad_norm": 2.069392204284668, |
|
"learning_rate": 4.601493301011534e-05, |
|
"loss": 0.0228, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.8984744871120462, |
|
"grad_norm": 1.4139840602874756, |
|
"learning_rate": 4.6007076066874786e-05, |
|
"loss": 0.016, |
|
"step": 1281 |
|
}, |
|
{ |
|
"epoch": 0.8991758723478871, |
|
"grad_norm": 0.8678575158119202, |
|
"learning_rate": 4.599921205805663e-05, |
|
"loss": 0.0092, |
|
"step": 1282 |
|
}, |
|
{ |
|
"epoch": 0.8998772575837278, |
|
"grad_norm": 0.3188234567642212, |
|
"learning_rate": 4.5991340986305896e-05, |
|
"loss": 0.0031, |
|
"step": 1283 |
|
}, |
|
{ |
|
"epoch": 0.9005786428195687, |
|
"grad_norm": 0.13479161262512207, |
|
"learning_rate": 4.598346285426994e-05, |
|
"loss": 0.0012, |
|
"step": 1284 |
|
}, |
|
{ |
|
"epoch": 0.9012800280554094, |
|
"grad_norm": 44.85018539428711, |
|
"learning_rate": 4.5975577664598535e-05, |
|
"loss": 1.8893, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.9019814132912503, |
|
"grad_norm": 0.030871881172060966, |
|
"learning_rate": 4.596768541994379e-05, |
|
"loss": 0.0003, |
|
"step": 1286 |
|
}, |
|
{ |
|
"epoch": 0.902682798527091, |
|
"grad_norm": 0.025855490937829018, |
|
"learning_rate": 4.595978612296021e-05, |
|
"loss": 0.0002, |
|
"step": 1287 |
|
}, |
|
{ |
|
"epoch": 0.9033841837629318, |
|
"grad_norm": 46.28804397583008, |
|
"learning_rate": 4.595187977630467e-05, |
|
"loss": 2.4086, |
|
"step": 1288 |
|
}, |
|
{ |
|
"epoch": 0.9040855689987726, |
|
"grad_norm": 46.17903518676758, |
|
"learning_rate": 4.5943966382636413e-05, |
|
"loss": 2.5441, |
|
"step": 1289 |
|
}, |
|
{ |
|
"epoch": 0.9047869542346134, |
|
"grad_norm": 48.56827926635742, |
|
"learning_rate": 4.593604594461704e-05, |
|
"loss": 2.5474, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.9054883394704542, |
|
"grad_norm": 47.86493682861328, |
|
"learning_rate": 4.592811846491055e-05, |
|
"loss": 2.1781, |
|
"step": 1291 |
|
}, |
|
{ |
|
"epoch": 0.9061897247062949, |
|
"grad_norm": 43.444766998291016, |
|
"learning_rate": 4.592018394618328e-05, |
|
"loss": 1.8371, |
|
"step": 1292 |
|
}, |
|
{ |
|
"epoch": 0.9068911099421357, |
|
"grad_norm": 46.283565521240234, |
|
"learning_rate": 4.5912242391103944e-05, |
|
"loss": 1.5648, |
|
"step": 1293 |
|
}, |
|
{ |
|
"epoch": 0.9075924951779765, |
|
"grad_norm": 0.2855938971042633, |
|
"learning_rate": 4.5904293802343646e-05, |
|
"loss": 0.0028, |
|
"step": 1294 |
|
}, |
|
{ |
|
"epoch": 0.9082938804138173, |
|
"grad_norm": 44.066627502441406, |
|
"learning_rate": 4.589633818257582e-05, |
|
"loss": 1.3621, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.908995265649658, |
|
"grad_norm": 1.1252837181091309, |
|
"learning_rate": 4.588837553447628e-05, |
|
"loss": 0.0122, |
|
"step": 1296 |
|
}, |
|
{ |
|
"epoch": 0.9096966508854989, |
|
"grad_norm": 1.7556012868881226, |
|
"learning_rate": 4.588040586072321e-05, |
|
"loss": 0.0199, |
|
"step": 1297 |
|
}, |
|
{ |
|
"epoch": 0.9103980361213396, |
|
"grad_norm": 2.0839264392852783, |
|
"learning_rate": 4.5872429163997163e-05, |
|
"loss": 0.0248, |
|
"step": 1298 |
|
}, |
|
{ |
|
"epoch": 0.9110994213571805, |
|
"grad_norm": 2.153773784637451, |
|
"learning_rate": 4.5864445446981025e-05, |
|
"loss": 0.0254, |
|
"step": 1299 |
|
}, |
|
{ |
|
"epoch": 0.9118008065930212, |
|
"grad_norm": 2.1063156127929688, |
|
"learning_rate": 4.5856454712360067e-05, |
|
"loss": 0.0235, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.912502191828862, |
|
"grad_norm": 1.26010000705719, |
|
"learning_rate": 4.584845696282192e-05, |
|
"loss": 0.0134, |
|
"step": 1301 |
|
}, |
|
{ |
|
"epoch": 0.9132035770647028, |
|
"grad_norm": 0.7219848036766052, |
|
"learning_rate": 4.584045220105656e-05, |
|
"loss": 0.0075, |
|
"step": 1302 |
|
}, |
|
{ |
|
"epoch": 0.9139049623005435, |
|
"grad_norm": 0.3715592324733734, |
|
"learning_rate": 4.583244042975634e-05, |
|
"loss": 0.0037, |
|
"step": 1303 |
|
}, |
|
{ |
|
"epoch": 0.9146063475363844, |
|
"grad_norm": 0.16978369653224945, |
|
"learning_rate": 4.582442165161596e-05, |
|
"loss": 0.0017, |
|
"step": 1304 |
|
}, |
|
{ |
|
"epoch": 0.9153077327722251, |
|
"grad_norm": 0.10813824832439423, |
|
"learning_rate": 4.5816395869332476e-05, |
|
"loss": 0.001, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.916009118008066, |
|
"grad_norm": 0.05026814341545105, |
|
"learning_rate": 4.5808363085605296e-05, |
|
"loss": 0.0005, |
|
"step": 1306 |
|
}, |
|
{ |
|
"epoch": 0.9167105032439067, |
|
"grad_norm": 0.025502556934952736, |
|
"learning_rate": 4.58003233031362e-05, |
|
"loss": 0.0002, |
|
"step": 1307 |
|
}, |
|
{ |
|
"epoch": 0.9174118884797475, |
|
"grad_norm": 45.29407501220703, |
|
"learning_rate": 4.579227652462928e-05, |
|
"loss": 2.3544, |
|
"step": 1308 |
|
}, |
|
{ |
|
"epoch": 0.9181132737155883, |
|
"grad_norm": 0.012745014391839504, |
|
"learning_rate": 4.578422275279105e-05, |
|
"loss": 0.0001, |
|
"step": 1309 |
|
}, |
|
{ |
|
"epoch": 0.9188146589514291, |
|
"grad_norm": 0.009426114149391651, |
|
"learning_rate": 4.577616199033033e-05, |
|
"loss": 0.0001, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.9195160441872698, |
|
"grad_norm": 0.008077832870185375, |
|
"learning_rate": 4.5768094239958285e-05, |
|
"loss": 0.0001, |
|
"step": 1311 |
|
}, |
|
{ |
|
"epoch": 0.9202174294231107, |
|
"grad_norm": 0.006128115579485893, |
|
"learning_rate": 4.576001950438845e-05, |
|
"loss": 0.0001, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 0.9209188146589514, |
|
"grad_norm": 88.39704895019531, |
|
"learning_rate": 4.575193778633671e-05, |
|
"loss": 5.5453, |
|
"step": 1313 |
|
}, |
|
{ |
|
"epoch": 0.9216201998947923, |
|
"grad_norm": 45.73112869262695, |
|
"learning_rate": 4.574384908852128e-05, |
|
"loss": 2.724, |
|
"step": 1314 |
|
}, |
|
{ |
|
"epoch": 0.922321585130633, |
|
"grad_norm": 45.08832550048828, |
|
"learning_rate": 4.573575341366274e-05, |
|
"loss": 2.5745, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.9230229703664737, |
|
"grad_norm": 0.01642024703323841, |
|
"learning_rate": 4.572765076448402e-05, |
|
"loss": 0.0001, |
|
"step": 1316 |
|
}, |
|
{ |
|
"epoch": 0.9237243556023146, |
|
"grad_norm": 0.0225964467972517, |
|
"learning_rate": 4.5719541143710375e-05, |
|
"loss": 0.0002, |
|
"step": 1317 |
|
}, |
|
{ |
|
"epoch": 0.9244257408381553, |
|
"grad_norm": 0.03809621185064316, |
|
"learning_rate": 4.571142455406943e-05, |
|
"loss": 0.0003, |
|
"step": 1318 |
|
}, |
|
{ |
|
"epoch": 0.9251271260739962, |
|
"grad_norm": 42.60374450683594, |
|
"learning_rate": 4.5703300998291124e-05, |
|
"loss": 1.805, |
|
"step": 1319 |
|
}, |
|
{ |
|
"epoch": 0.9258285113098369, |
|
"grad_norm": 0.08730126917362213, |
|
"learning_rate": 4.569517047910777e-05, |
|
"loss": 0.0008, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.9265298965456777, |
|
"grad_norm": 119.11152648925781, |
|
"learning_rate": 4.5687032999254e-05, |
|
"loss": 4.0878, |
|
"step": 1321 |
|
}, |
|
{ |
|
"epoch": 0.9272312817815185, |
|
"grad_norm": 0.3243617117404938, |
|
"learning_rate": 4.567888856146681e-05, |
|
"loss": 0.0031, |
|
"step": 1322 |
|
}, |
|
{ |
|
"epoch": 0.9279326670173593, |
|
"grad_norm": 0.4152386784553528, |
|
"learning_rate": 4.567073716848551e-05, |
|
"loss": 0.004, |
|
"step": 1323 |
|
}, |
|
{ |
|
"epoch": 0.9286340522532001, |
|
"grad_norm": 30.812305450439453, |
|
"learning_rate": 4.566257882305176e-05, |
|
"loss": 0.7747, |
|
"step": 1324 |
|
}, |
|
{ |
|
"epoch": 0.9293354374890409, |
|
"grad_norm": 30.550561904907227, |
|
"learning_rate": 4.565441352790957e-05, |
|
"loss": 0.7148, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.9300368227248816, |
|
"grad_norm": 79.6876220703125, |
|
"learning_rate": 4.564624128580528e-05, |
|
"loss": 1.728, |
|
"step": 1326 |
|
}, |
|
{ |
|
"epoch": 0.9307382079607224, |
|
"grad_norm": 3.2627832889556885, |
|
"learning_rate": 4.5638062099487545e-05, |
|
"loss": 0.0427, |
|
"step": 1327 |
|
}, |
|
{ |
|
"epoch": 0.9314395931965632, |
|
"grad_norm": 10.727757453918457, |
|
"learning_rate": 4.5629875971707395e-05, |
|
"loss": 0.2789, |
|
"step": 1328 |
|
}, |
|
{ |
|
"epoch": 0.932140978432404, |
|
"grad_norm": 5.946654319763184, |
|
"learning_rate": 4.562168290521817e-05, |
|
"loss": 0.0887, |
|
"step": 1329 |
|
}, |
|
{ |
|
"epoch": 0.9328423636682448, |
|
"grad_norm": 1.3037784099578857, |
|
"learning_rate": 4.561348290277554e-05, |
|
"loss": 0.1909, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.9335437489040855, |
|
"grad_norm": 8.197032928466797, |
|
"learning_rate": 4.560527596713752e-05, |
|
"loss": 0.1388, |
|
"step": 1331 |
|
}, |
|
{ |
|
"epoch": 0.9342451341399264, |
|
"grad_norm": 8.003156661987305, |
|
"learning_rate": 4.559706210106446e-05, |
|
"loss": 0.1299, |
|
"step": 1332 |
|
}, |
|
{ |
|
"epoch": 0.9349465193757671, |
|
"grad_norm": 13.647798538208008, |
|
"learning_rate": 4.558884130731903e-05, |
|
"loss": 0.3574, |
|
"step": 1333 |
|
}, |
|
{ |
|
"epoch": 0.935647904611608, |
|
"grad_norm": 5.755825996398926, |
|
"learning_rate": 4.558061358866623e-05, |
|
"loss": 0.0832, |
|
"step": 1334 |
|
}, |
|
{ |
|
"epoch": 0.9363492898474487, |
|
"grad_norm": 21.87740135192871, |
|
"learning_rate": 4.55723789478734e-05, |
|
"loss": 0.4666, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.9370506750832895, |
|
"grad_norm": 9.851075172424316, |
|
"learning_rate": 4.5564137387710196e-05, |
|
"loss": 0.2624, |
|
"step": 1336 |
|
}, |
|
{ |
|
"epoch": 0.9377520603191303, |
|
"grad_norm": 4.983445644378662, |
|
"learning_rate": 4.555588891094862e-05, |
|
"loss": 0.0688, |
|
"step": 1337 |
|
}, |
|
{ |
|
"epoch": 0.9384534455549711, |
|
"grad_norm": 4.145101547241211, |
|
"learning_rate": 4.554763352036296e-05, |
|
"loss": 0.0537, |
|
"step": 1338 |
|
}, |
|
{ |
|
"epoch": 0.9391548307908119, |
|
"grad_norm": 4.348505020141602, |
|
"learning_rate": 4.5539371218729894e-05, |
|
"loss": 0.058, |
|
"step": 1339 |
|
}, |
|
{ |
|
"epoch": 0.9398562160266526, |
|
"grad_norm": 16.51106834411621, |
|
"learning_rate": 4.5531102008828354e-05, |
|
"loss": 0.3685, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.9405576012624934, |
|
"grad_norm": 2.958770513534546, |
|
"learning_rate": 4.5522825893439645e-05, |
|
"loss": 0.0375, |
|
"step": 1341 |
|
}, |
|
{ |
|
"epoch": 0.9412589864983342, |
|
"grad_norm": 1.705260157585144, |
|
"learning_rate": 4.5514542875347376e-05, |
|
"loss": 0.0194, |
|
"step": 1342 |
|
}, |
|
{ |
|
"epoch": 0.941960371734175, |
|
"grad_norm": 28.18739891052246, |
|
"learning_rate": 4.550625295733748e-05, |
|
"loss": 0.6777, |
|
"step": 1343 |
|
}, |
|
{ |
|
"epoch": 0.9426617569700158, |
|
"grad_norm": 39.53862380981445, |
|
"learning_rate": 4.5497956142198214e-05, |
|
"loss": 1.1208, |
|
"step": 1344 |
|
}, |
|
{ |
|
"epoch": 0.9433631422058566, |
|
"grad_norm": 1.0771280527114868, |
|
"learning_rate": 4.5489652432720145e-05, |
|
"loss": 0.0118, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.9440645274416973, |
|
"grad_norm": 96.40424346923828, |
|
"learning_rate": 4.548134183169617e-05, |
|
"loss": 2.4673, |
|
"step": 1346 |
|
}, |
|
{ |
|
"epoch": 0.9447659126775382, |
|
"grad_norm": 0.5970985293388367, |
|
"learning_rate": 4.547302434192149e-05, |
|
"loss": 0.0061, |
|
"step": 1347 |
|
}, |
|
{ |
|
"epoch": 0.9454672979133789, |
|
"grad_norm": 36.97736358642578, |
|
"learning_rate": 4.546469996619364e-05, |
|
"loss": 0.9705, |
|
"step": 1348 |
|
}, |
|
{ |
|
"epoch": 0.9461686831492198, |
|
"grad_norm": 0.6519697308540344, |
|
"learning_rate": 4.545636870731247e-05, |
|
"loss": 0.0066, |
|
"step": 1349 |
|
}, |
|
{ |
|
"epoch": 0.9468700683850605, |
|
"grad_norm": 0.4174421429634094, |
|
"learning_rate": 4.544803056808013e-05, |
|
"loss": 0.0042, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.9475714536209012, |
|
"grad_norm": 0.4273722469806671, |
|
"learning_rate": 4.5439685551301094e-05, |
|
"loss": 0.0043, |
|
"step": 1351 |
|
}, |
|
{ |
|
"epoch": 0.9482728388567421, |
|
"grad_norm": 0.3083280324935913, |
|
"learning_rate": 4.543133365978214e-05, |
|
"loss": 0.003, |
|
"step": 1352 |
|
}, |
|
{ |
|
"epoch": 0.9489742240925828, |
|
"grad_norm": 0.2918597161769867, |
|
"learning_rate": 4.542297489633238e-05, |
|
"loss": 0.0028, |
|
"step": 1353 |
|
}, |
|
{ |
|
"epoch": 0.9496756093284237, |
|
"grad_norm": 47.360557556152344, |
|
"learning_rate": 4.541460926376322e-05, |
|
"loss": 1.7171, |
|
"step": 1354 |
|
}, |
|
{ |
|
"epoch": 0.9503769945642644, |
|
"grad_norm": 81.7466049194336, |
|
"learning_rate": 4.5406236764888374e-05, |
|
"loss": 2.8326, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.9510783798001052, |
|
"grad_norm": 39.265384674072266, |
|
"learning_rate": 4.539785740252388e-05, |
|
"loss": 1.3337, |
|
"step": 1356 |
|
}, |
|
{ |
|
"epoch": 0.951779765035946, |
|
"grad_norm": 0.28257325291633606, |
|
"learning_rate": 4.5389471179488074e-05, |
|
"loss": 0.0027, |
|
"step": 1357 |
|
}, |
|
{ |
|
"epoch": 0.9524811502717868, |
|
"grad_norm": 41.259849548339844, |
|
"learning_rate": 4.538107809860159e-05, |
|
"loss": 1.2651, |
|
"step": 1358 |
|
}, |
|
{ |
|
"epoch": 0.9531825355076275, |
|
"grad_norm": 0.5483337640762329, |
|
"learning_rate": 4.537267816268739e-05, |
|
"loss": 0.0055, |
|
"step": 1359 |
|
}, |
|
{ |
|
"epoch": 0.9538839207434684, |
|
"grad_norm": 0.9649966359138489, |
|
"learning_rate": 4.536427137457072e-05, |
|
"loss": 0.0098, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.9545853059793091, |
|
"grad_norm": 28.19986915588379, |
|
"learning_rate": 4.5355857737079166e-05, |
|
"loss": 0.6184, |
|
"step": 1361 |
|
}, |
|
{ |
|
"epoch": 0.95528669121515, |
|
"grad_norm": 1.310431957244873, |
|
"learning_rate": 4.534743725304257e-05, |
|
"loss": 0.0142, |
|
"step": 1362 |
|
}, |
|
{ |
|
"epoch": 0.9559880764509907, |
|
"grad_norm": 1.170506477355957, |
|
"learning_rate": 4.5339009925293105e-05, |
|
"loss": 0.0125, |
|
"step": 1363 |
|
}, |
|
{ |
|
"epoch": 0.9566894616868314, |
|
"grad_norm": 1.135798454284668, |
|
"learning_rate": 4.533057575666526e-05, |
|
"loss": 0.0118, |
|
"step": 1364 |
|
}, |
|
{ |
|
"epoch": 0.9573908469226723, |
|
"grad_norm": 0.7438721060752869, |
|
"learning_rate": 4.532213474999578e-05, |
|
"loss": 0.0077, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.958092232158513, |
|
"grad_norm": 0.46775007247924805, |
|
"learning_rate": 4.531368690812376e-05, |
|
"loss": 0.0046, |
|
"step": 1366 |
|
}, |
|
{ |
|
"epoch": 0.9587936173943539, |
|
"grad_norm": 0.2829347848892212, |
|
"learning_rate": 4.530523223389056e-05, |
|
"loss": 0.0027, |
|
"step": 1367 |
|
}, |
|
{ |
|
"epoch": 0.9594950026301946, |
|
"grad_norm": 0.12059792876243591, |
|
"learning_rate": 4.529677073013984e-05, |
|
"loss": 0.0011, |
|
"step": 1368 |
|
}, |
|
{ |
|
"epoch": 0.9601963878660355, |
|
"grad_norm": 41.76416015625, |
|
"learning_rate": 4.5288302399717575e-05, |
|
"loss": 1.5052, |
|
"step": 1369 |
|
}, |
|
{ |
|
"epoch": 0.9608977731018762, |
|
"grad_norm": 0.05551628768444061, |
|
"learning_rate": 4.5279827245472025e-05, |
|
"loss": 0.0005, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.961599158337717, |
|
"grad_norm": 44.579280853271484, |
|
"learning_rate": 4.527134527025374e-05, |
|
"loss": 1.9268, |
|
"step": 1371 |
|
}, |
|
{ |
|
"epoch": 0.9623005435735578, |
|
"grad_norm": 92.77808380126953, |
|
"learning_rate": 4.526285647691559e-05, |
|
"loss": 3.8084, |
|
"step": 1372 |
|
}, |
|
{ |
|
"epoch": 0.9630019288093986, |
|
"grad_norm": 0.06515678018331528, |
|
"learning_rate": 4.5254360868312694e-05, |
|
"loss": 0.0006, |
|
"step": 1373 |
|
}, |
|
{ |
|
"epoch": 0.9637033140452393, |
|
"grad_norm": 44.31393814086914, |
|
"learning_rate": 4.5245858447302506e-05, |
|
"loss": 1.8936, |
|
"step": 1374 |
|
}, |
|
{ |
|
"epoch": 0.9644046992810801, |
|
"grad_norm": 42.123191833496094, |
|
"learning_rate": 4.523734921674475e-05, |
|
"loss": 1.5489, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.9651060845169209, |
|
"grad_norm": 78.27152252197266, |
|
"learning_rate": 4.522883317950144e-05, |
|
"loss": 1.4444, |
|
"step": 1376 |
|
}, |
|
{ |
|
"epoch": 0.9658074697527617, |
|
"grad_norm": 0.4181966185569763, |
|
"learning_rate": 4.5220310338436885e-05, |
|
"loss": 0.0041, |
|
"step": 1377 |
|
}, |
|
{ |
|
"epoch": 0.9665088549886025, |
|
"grad_norm": 61.79771423339844, |
|
"learning_rate": 4.521178069641768e-05, |
|
"loss": 1.5033, |
|
"step": 1378 |
|
}, |
|
{ |
|
"epoch": 0.9672102402244432, |
|
"grad_norm": 1.026915431022644, |
|
"learning_rate": 4.520324425631271e-05, |
|
"loss": 0.011, |
|
"step": 1379 |
|
}, |
|
{ |
|
"epoch": 0.9679116254602841, |
|
"grad_norm": 1.1025463342666626, |
|
"learning_rate": 4.519470102099315e-05, |
|
"loss": 0.0117, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.9686130106961248, |
|
"grad_norm": 1.5587780475616455, |
|
"learning_rate": 4.518615099333243e-05, |
|
"loss": 0.018, |
|
"step": 1381 |
|
}, |
|
{ |
|
"epoch": 0.9693143959319657, |
|
"grad_norm": 1.4015796184539795, |
|
"learning_rate": 4.517759417620633e-05, |
|
"loss": 0.016, |
|
"step": 1382 |
|
}, |
|
{ |
|
"epoch": 0.9700157811678064, |
|
"grad_norm": 1.1269611120224, |
|
"learning_rate": 4.516903057249284e-05, |
|
"loss": 0.0124, |
|
"step": 1383 |
|
}, |
|
{ |
|
"epoch": 0.9707171664036472, |
|
"grad_norm": 64.7952651977539, |
|
"learning_rate": 4.516046018507229e-05, |
|
"loss": 1.7053, |
|
"step": 1384 |
|
}, |
|
{ |
|
"epoch": 0.971418551639488, |
|
"grad_norm": 36.87698745727539, |
|
"learning_rate": 4.515188301682725e-05, |
|
"loss": 0.8805, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.9721199368753288, |
|
"grad_norm": 39.0102653503418, |
|
"learning_rate": 4.5143299070642584e-05, |
|
"loss": 1.2212, |
|
"step": 1386 |
|
}, |
|
{ |
|
"epoch": 0.9728213221111696, |
|
"grad_norm": 62.35078811645508, |
|
"learning_rate": 4.5134708349405457e-05, |
|
"loss": 1.6357, |
|
"step": 1387 |
|
}, |
|
{ |
|
"epoch": 0.9735227073470103, |
|
"grad_norm": 31.161195755004883, |
|
"learning_rate": 4.5126110856005285e-05, |
|
"loss": 0.7379, |
|
"step": 1388 |
|
}, |
|
{ |
|
"epoch": 0.9742240925828511, |
|
"grad_norm": 1.4460724592208862, |
|
"learning_rate": 4.511750659333379e-05, |
|
"loss": 0.0166, |
|
"step": 1389 |
|
}, |
|
{ |
|
"epoch": 0.9749254778186919, |
|
"grad_norm": 1.6871577501296997, |
|
"learning_rate": 4.5108895564284924e-05, |
|
"loss": 0.0201, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.9756268630545327, |
|
"grad_norm": 18.601655960083008, |
|
"learning_rate": 4.510027777175497e-05, |
|
"loss": 0.3773, |
|
"step": 1391 |
|
}, |
|
{ |
|
"epoch": 0.9763282482903735, |
|
"grad_norm": 24.591615676879883, |
|
"learning_rate": 4.509165321864245e-05, |
|
"loss": 0.6012, |
|
"step": 1392 |
|
}, |
|
{ |
|
"epoch": 0.9770296335262143, |
|
"grad_norm": 2.3374860286712646, |
|
"learning_rate": 4.5083021907848164e-05, |
|
"loss": 0.0286, |
|
"step": 1393 |
|
}, |
|
{ |
|
"epoch": 0.977731018762055, |
|
"grad_norm": 3.1211438179016113, |
|
"learning_rate": 4.50743838422752e-05, |
|
"loss": 0.0388, |
|
"step": 1394 |
|
}, |
|
{ |
|
"epoch": 0.9784324039978959, |
|
"grad_norm": 15.115840911865234, |
|
"learning_rate": 4.50657390248289e-05, |
|
"loss": 0.3451, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.9791337892337366, |
|
"grad_norm": 13.081830978393555, |
|
"learning_rate": 4.50570874584169e-05, |
|
"loss": 0.3082, |
|
"step": 1396 |
|
}, |
|
{ |
|
"epoch": 0.9798351744695775, |
|
"grad_norm": 3.7496583461761475, |
|
"learning_rate": 4.5048429145949076e-05, |
|
"loss": 0.0503, |
|
"step": 1397 |
|
}, |
|
{ |
|
"epoch": 0.9805365597054182, |
|
"grad_norm": 22.880462646484375, |
|
"learning_rate": 4.503976409033759e-05, |
|
"loss": 0.4432, |
|
"step": 1398 |
|
}, |
|
{ |
|
"epoch": 0.9812379449412589, |
|
"grad_norm": 4.944215774536133, |
|
"learning_rate": 4.503109229449688e-05, |
|
"loss": 0.0706, |
|
"step": 1399 |
|
}, |
|
{ |
|
"epoch": 0.9819393301770998, |
|
"grad_norm": 4.719092845916748, |
|
"learning_rate": 4.502241376134364e-05, |
|
"loss": 0.0667, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.9826407154129405, |
|
"grad_norm": 4.378533363342285, |
|
"learning_rate": 4.5013728493796826e-05, |
|
"loss": 0.0611, |
|
"step": 1401 |
|
}, |
|
{ |
|
"epoch": 0.9833421006487814, |
|
"grad_norm": 4.234534740447998, |
|
"learning_rate": 4.5005036494777666e-05, |
|
"loss": 0.055, |
|
"step": 1402 |
|
}, |
|
{ |
|
"epoch": 0.9840434858846221, |
|
"grad_norm": 2.942084789276123, |
|
"learning_rate": 4.4996337767209656e-05, |
|
"loss": 0.0382, |
|
"step": 1403 |
|
}, |
|
{ |
|
"epoch": 0.9847448711204629, |
|
"grad_norm": 1.5875838994979858, |
|
"learning_rate": 4.498763231401855e-05, |
|
"loss": 0.018, |
|
"step": 1404 |
|
}, |
|
{ |
|
"epoch": 0.9854462563563037, |
|
"grad_norm": 53.9989128112793, |
|
"learning_rate": 4.497892013813235e-05, |
|
"loss": 1.2552, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.9861476415921445, |
|
"grad_norm": 28.49068832397461, |
|
"learning_rate": 4.497020124248135e-05, |
|
"loss": 0.7336, |
|
"step": 1406 |
|
}, |
|
{ |
|
"epoch": 0.9868490268279853, |
|
"grad_norm": 0.4417467713356018, |
|
"learning_rate": 4.496147562999807e-05, |
|
"loss": 0.0045, |
|
"step": 1407 |
|
}, |
|
{ |
|
"epoch": 0.9875504120638261, |
|
"grad_norm": 104.99906921386719, |
|
"learning_rate": 4.4952743303617325e-05, |
|
"loss": 3.0756, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 0.9882517972996668, |
|
"grad_norm": 0.3293454647064209, |
|
"learning_rate": 4.4944004266276164e-05, |
|
"loss": 0.0033, |
|
"step": 1409 |
|
}, |
|
{ |
|
"epoch": 0.9889531825355077, |
|
"grad_norm": 41.633243560791016, |
|
"learning_rate": 4.4935258520913894e-05, |
|
"loss": 1.3174, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.9896545677713484, |
|
"grad_norm": 0.23380160331726074, |
|
"learning_rate": 4.4926506070472085e-05, |
|
"loss": 0.0023, |
|
"step": 1411 |
|
}, |
|
{ |
|
"epoch": 0.9903559530071891, |
|
"grad_norm": 0.44888490438461304, |
|
"learning_rate": 4.4917746917894556e-05, |
|
"loss": 0.0047, |
|
"step": 1412 |
|
}, |
|
{ |
|
"epoch": 0.99105733824303, |
|
"grad_norm": 33.958248138427734, |
|
"learning_rate": 4.490898106612739e-05, |
|
"loss": 0.9667, |
|
"step": 1413 |
|
}, |
|
{ |
|
"epoch": 0.9917587234788707, |
|
"grad_norm": 0.3081722855567932, |
|
"learning_rate": 4.490020851811892e-05, |
|
"loss": 0.003, |
|
"step": 1414 |
|
}, |
|
{ |
|
"epoch": 0.9924601087147116, |
|
"grad_norm": 32.04133224487305, |
|
"learning_rate": 4.4891429276819726e-05, |
|
"loss": 0.9003, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.9931614939505523, |
|
"grad_norm": 43.6740608215332, |
|
"learning_rate": 4.488264334518264e-05, |
|
"loss": 0.9288, |
|
"step": 1416 |
|
}, |
|
{ |
|
"epoch": 0.9938628791863932, |
|
"grad_norm": 29.450580596923828, |
|
"learning_rate": 4.487385072616275e-05, |
|
"loss": 0.7131, |
|
"step": 1417 |
|
}, |
|
{ |
|
"epoch": 0.9945642644222339, |
|
"grad_norm": 1.1718950271606445, |
|
"learning_rate": 4.486505142271739e-05, |
|
"loss": 0.0137, |
|
"step": 1418 |
|
}, |
|
{ |
|
"epoch": 0.9952656496580747, |
|
"grad_norm": 1.5927923917770386, |
|
"learning_rate": 4.485624543780613e-05, |
|
"loss": 0.0192, |
|
"step": 1419 |
|
}, |
|
{ |
|
"epoch": 0.9959670348939155, |
|
"grad_norm": 24.835670471191406, |
|
"learning_rate": 4.4847432774390807e-05, |
|
"loss": 0.5717, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.9966684201297563, |
|
"grad_norm": 1.8400554656982422, |
|
"learning_rate": 4.4838613435435505e-05, |
|
"loss": 0.022, |
|
"step": 1421 |
|
}, |
|
{ |
|
"epoch": 0.997369805365597, |
|
"grad_norm": 2.3654353618621826, |
|
"learning_rate": 4.482978742390654e-05, |
|
"loss": 0.0298, |
|
"step": 1422 |
|
}, |
|
{ |
|
"epoch": 0.9980711906014378, |
|
"grad_norm": 24.44013023376465, |
|
"learning_rate": 4.482095474277247e-05, |
|
"loss": 0.5608, |
|
"step": 1423 |
|
}, |
|
{ |
|
"epoch": 0.9987725758372786, |
|
"grad_norm": 2.5648183822631836, |
|
"learning_rate": 4.48121153950041e-05, |
|
"loss": 0.034, |
|
"step": 1424 |
|
}, |
|
{ |
|
"epoch": 0.9994739610731194, |
|
"grad_norm": 2.0194385051727295, |
|
"learning_rate": 4.4803269383574507e-05, |
|
"loss": 0.0249, |
|
"step": 1425 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 1425, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.863941975950848e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|