|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 0, |
|
"global_step": 1274, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0007849293563579278, |
|
"grad_norm": 0.3427978456020355, |
|
"learning_rate": 1e-05, |
|
"loss": 2.1035, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0015698587127158557, |
|
"grad_norm": 0.332590788602829, |
|
"learning_rate": 9.992150706436422e-06, |
|
"loss": 2.0738, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.002354788069073783, |
|
"grad_norm": 0.36298516392707825, |
|
"learning_rate": 9.984301412872842e-06, |
|
"loss": 2.0758, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0031397174254317113, |
|
"grad_norm": 0.32631000876426697, |
|
"learning_rate": 9.976452119309263e-06, |
|
"loss": 1.9876, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.003924646781789639, |
|
"grad_norm": 0.36786606907844543, |
|
"learning_rate": 9.968602825745683e-06, |
|
"loss": 2.1213, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.004709576138147566, |
|
"grad_norm": 0.34687480330467224, |
|
"learning_rate": 9.960753532182104e-06, |
|
"loss": 2.0017, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.005494505494505495, |
|
"grad_norm": 0.3398713171482086, |
|
"learning_rate": 9.952904238618524e-06, |
|
"loss": 1.9805, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.006279434850863423, |
|
"grad_norm": 0.3185282051563263, |
|
"learning_rate": 9.945054945054946e-06, |
|
"loss": 1.8601, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.00706436420722135, |
|
"grad_norm": 0.372031033039093, |
|
"learning_rate": 9.937205651491367e-06, |
|
"loss": 2.1202, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.007849293563579277, |
|
"grad_norm": 0.3446281850337982, |
|
"learning_rate": 9.929356357927787e-06, |
|
"loss": 1.9856, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.008634222919937205, |
|
"grad_norm": 0.32924437522888184, |
|
"learning_rate": 9.921507064364208e-06, |
|
"loss": 1.8849, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.009419152276295133, |
|
"grad_norm": 0.33408552408218384, |
|
"learning_rate": 9.91365777080063e-06, |
|
"loss": 1.8825, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.01020408163265306, |
|
"grad_norm": 0.3477867841720581, |
|
"learning_rate": 9.90580847723705e-06, |
|
"loss": 1.9898, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.01098901098901099, |
|
"grad_norm": 0.3446689546108246, |
|
"learning_rate": 9.89795918367347e-06, |
|
"loss": 1.8995, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.011773940345368918, |
|
"grad_norm": 0.32618626952171326, |
|
"learning_rate": 9.890109890109892e-06, |
|
"loss": 1.8615, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.012558869701726845, |
|
"grad_norm": 0.3038957715034485, |
|
"learning_rate": 9.882260596546312e-06, |
|
"loss": 1.8644, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.013343799058084773, |
|
"grad_norm": 0.2970006465911865, |
|
"learning_rate": 9.874411302982733e-06, |
|
"loss": 1.8248, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0141287284144427, |
|
"grad_norm": 0.2924885153770447, |
|
"learning_rate": 9.866562009419153e-06, |
|
"loss": 1.8896, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.014913657770800628, |
|
"grad_norm": 0.30948498845100403, |
|
"learning_rate": 9.858712715855574e-06, |
|
"loss": 2.0174, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.015698587127158554, |
|
"grad_norm": 0.27170419692993164, |
|
"learning_rate": 9.850863422291994e-06, |
|
"loss": 1.8595, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.016483516483516484, |
|
"grad_norm": 0.2529163360595703, |
|
"learning_rate": 9.843014128728415e-06, |
|
"loss": 1.8447, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.01726844583987441, |
|
"grad_norm": 0.26019564270973206, |
|
"learning_rate": 9.835164835164835e-06, |
|
"loss": 1.8919, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.01805337519623234, |
|
"grad_norm": 0.251017302274704, |
|
"learning_rate": 9.827315541601256e-06, |
|
"loss": 1.8662, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.018838304552590265, |
|
"grad_norm": 0.22893306612968445, |
|
"learning_rate": 9.819466248037678e-06, |
|
"loss": 1.7685, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.019623233908948195, |
|
"grad_norm": 0.22592955827713013, |
|
"learning_rate": 9.811616954474098e-06, |
|
"loss": 1.7604, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.02040816326530612, |
|
"grad_norm": 0.22802409529685974, |
|
"learning_rate": 9.803767660910519e-06, |
|
"loss": 1.8441, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.02119309262166405, |
|
"grad_norm": 0.21545954048633575, |
|
"learning_rate": 9.795918367346939e-06, |
|
"loss": 1.7258, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.02197802197802198, |
|
"grad_norm": 0.22040924429893494, |
|
"learning_rate": 9.78806907378336e-06, |
|
"loss": 1.7892, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.022762951334379906, |
|
"grad_norm": 0.2393616884946823, |
|
"learning_rate": 9.780219780219781e-06, |
|
"loss": 1.8538, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.023547880690737835, |
|
"grad_norm": 0.20232702791690826, |
|
"learning_rate": 9.772370486656201e-06, |
|
"loss": 1.7536, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.02433281004709576, |
|
"grad_norm": 0.2203434556722641, |
|
"learning_rate": 9.764521193092623e-06, |
|
"loss": 1.7088, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.02511773940345369, |
|
"grad_norm": 0.1975688636302948, |
|
"learning_rate": 9.756671899529044e-06, |
|
"loss": 1.7131, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.025902668759811617, |
|
"grad_norm": 0.19861450791358948, |
|
"learning_rate": 9.748822605965464e-06, |
|
"loss": 1.776, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.026687598116169546, |
|
"grad_norm": 0.19722306728363037, |
|
"learning_rate": 9.740973312401885e-06, |
|
"loss": 1.6889, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.027472527472527472, |
|
"grad_norm": 0.18340197205543518, |
|
"learning_rate": 9.733124018838307e-06, |
|
"loss": 1.6312, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0282574568288854, |
|
"grad_norm": 0.20767861604690552, |
|
"learning_rate": 9.725274725274726e-06, |
|
"loss": 1.7293, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.029042386185243328, |
|
"grad_norm": 0.16855376958847046, |
|
"learning_rate": 9.717425431711148e-06, |
|
"loss": 1.5657, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.029827315541601257, |
|
"grad_norm": 0.19397929310798645, |
|
"learning_rate": 9.709576138147567e-06, |
|
"loss": 1.6843, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.030612244897959183, |
|
"grad_norm": 0.1965712457895279, |
|
"learning_rate": 9.701726844583989e-06, |
|
"loss": 1.7135, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.03139717425431711, |
|
"grad_norm": 0.18827113509178162, |
|
"learning_rate": 9.693877551020408e-06, |
|
"loss": 1.7002, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.03218210361067504, |
|
"grad_norm": 0.18982355296611786, |
|
"learning_rate": 9.68602825745683e-06, |
|
"loss": 1.6863, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.03296703296703297, |
|
"grad_norm": 0.18091963231563568, |
|
"learning_rate": 9.67817896389325e-06, |
|
"loss": 1.6784, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.033751962323390894, |
|
"grad_norm": 0.17879721522331238, |
|
"learning_rate": 9.670329670329671e-06, |
|
"loss": 1.6552, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.03453689167974882, |
|
"grad_norm": 0.19234947860240936, |
|
"learning_rate": 9.66248037676609e-06, |
|
"loss": 1.7857, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.03532182103610675, |
|
"grad_norm": 0.16198568046092987, |
|
"learning_rate": 9.654631083202512e-06, |
|
"loss": 1.5865, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.03610675039246468, |
|
"grad_norm": 0.16286341845989227, |
|
"learning_rate": 9.646781789638933e-06, |
|
"loss": 1.6461, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.036891679748822605, |
|
"grad_norm": 0.15107131004333496, |
|
"learning_rate": 9.638932496075353e-06, |
|
"loss": 1.6048, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.03767660910518053, |
|
"grad_norm": 0.16922970116138458, |
|
"learning_rate": 9.631083202511775e-06, |
|
"loss": 1.6644, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.038461538461538464, |
|
"grad_norm": 0.15843363106250763, |
|
"learning_rate": 9.623233908948196e-06, |
|
"loss": 1.5417, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.03924646781789639, |
|
"grad_norm": 0.17787528038024902, |
|
"learning_rate": 9.615384615384616e-06, |
|
"loss": 1.6869, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.040031397174254316, |
|
"grad_norm": 0.1719801276922226, |
|
"learning_rate": 9.607535321821037e-06, |
|
"loss": 1.6351, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.04081632653061224, |
|
"grad_norm": 0.1807231903076172, |
|
"learning_rate": 9.599686028257459e-06, |
|
"loss": 1.6697, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.041601255886970175, |
|
"grad_norm": 0.1894819140434265, |
|
"learning_rate": 9.591836734693878e-06, |
|
"loss": 1.6668, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.0423861852433281, |
|
"grad_norm": 0.16167454421520233, |
|
"learning_rate": 9.5839874411303e-06, |
|
"loss": 1.601, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.04317111459968603, |
|
"grad_norm": 0.15732350945472717, |
|
"learning_rate": 9.576138147566721e-06, |
|
"loss": 1.5291, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.04395604395604396, |
|
"grad_norm": 0.16725674271583557, |
|
"learning_rate": 9.56828885400314e-06, |
|
"loss": 1.6179, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.044740973312401885, |
|
"grad_norm": 0.1474035680294037, |
|
"learning_rate": 9.560439560439562e-06, |
|
"loss": 1.5644, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.04552590266875981, |
|
"grad_norm": 0.15364201366901398, |
|
"learning_rate": 9.552590266875982e-06, |
|
"loss": 1.6396, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.04631083202511774, |
|
"grad_norm": 0.19037100672721863, |
|
"learning_rate": 9.544740973312403e-06, |
|
"loss": 1.561, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.04709576138147567, |
|
"grad_norm": 0.14766716957092285, |
|
"learning_rate": 9.536891679748823e-06, |
|
"loss": 1.5349, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.047880690737833596, |
|
"grad_norm": 0.13931486010551453, |
|
"learning_rate": 9.529042386185244e-06, |
|
"loss": 1.4435, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.04866562009419152, |
|
"grad_norm": 0.1515316218137741, |
|
"learning_rate": 9.521193092621664e-06, |
|
"loss": 1.5309, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.04945054945054945, |
|
"grad_norm": 0.14389821887016296, |
|
"learning_rate": 9.513343799058085e-06, |
|
"loss": 1.4471, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.05023547880690738, |
|
"grad_norm": 0.14060775935649872, |
|
"learning_rate": 9.505494505494505e-06, |
|
"loss": 1.5074, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.05102040816326531, |
|
"grad_norm": 0.1455729603767395, |
|
"learning_rate": 9.497645211930927e-06, |
|
"loss": 1.5528, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.05180533751962323, |
|
"grad_norm": 0.14463943243026733, |
|
"learning_rate": 9.489795918367348e-06, |
|
"loss": 1.4809, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.05259026687598116, |
|
"grad_norm": 0.1522558331489563, |
|
"learning_rate": 9.481946624803768e-06, |
|
"loss": 1.4754, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.05337519623233909, |
|
"grad_norm": 0.1452169567346573, |
|
"learning_rate": 9.474097331240189e-06, |
|
"loss": 1.4867, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.05416012558869702, |
|
"grad_norm": 0.14983990788459778, |
|
"learning_rate": 9.46624803767661e-06, |
|
"loss": 1.4852, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.054945054945054944, |
|
"grad_norm": 0.15380005538463593, |
|
"learning_rate": 9.45839874411303e-06, |
|
"loss": 1.5105, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.05572998430141287, |
|
"grad_norm": 0.1510206013917923, |
|
"learning_rate": 9.450549450549452e-06, |
|
"loss": 1.5202, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.0565149136577708, |
|
"grad_norm": 0.1501176506280899, |
|
"learning_rate": 9.442700156985873e-06, |
|
"loss": 1.5119, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.05729984301412873, |
|
"grad_norm": 0.26818370819091797, |
|
"learning_rate": 9.434850863422293e-06, |
|
"loss": 1.5876, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.058084772370486655, |
|
"grad_norm": 0.1636001467704773, |
|
"learning_rate": 9.427001569858714e-06, |
|
"loss": 1.5482, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.05886970172684458, |
|
"grad_norm": 0.1485077142715454, |
|
"learning_rate": 9.419152276295134e-06, |
|
"loss": 1.5007, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.059654631083202514, |
|
"grad_norm": 0.14850012958049774, |
|
"learning_rate": 9.411302982731555e-06, |
|
"loss": 1.4424, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.06043956043956044, |
|
"grad_norm": 0.16702501475811005, |
|
"learning_rate": 9.403453689167977e-06, |
|
"loss": 1.5523, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.061224489795918366, |
|
"grad_norm": 0.15823277831077576, |
|
"learning_rate": 9.395604395604396e-06, |
|
"loss": 1.4958, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.06200941915227629, |
|
"grad_norm": 0.15339218080043793, |
|
"learning_rate": 9.387755102040818e-06, |
|
"loss": 1.4549, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.06279434850863422, |
|
"grad_norm": 0.15759187936782837, |
|
"learning_rate": 9.379905808477237e-06, |
|
"loss": 1.4164, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.06357927786499215, |
|
"grad_norm": 0.16237343847751617, |
|
"learning_rate": 9.372056514913659e-06, |
|
"loss": 1.5056, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.06436420722135008, |
|
"grad_norm": 0.145219624042511, |
|
"learning_rate": 9.364207221350079e-06, |
|
"loss": 1.4138, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.065149136577708, |
|
"grad_norm": 0.16004693508148193, |
|
"learning_rate": 9.3563579277865e-06, |
|
"loss": 1.4653, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.06593406593406594, |
|
"grad_norm": 0.17298448085784912, |
|
"learning_rate": 9.34850863422292e-06, |
|
"loss": 1.4232, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.06671899529042387, |
|
"grad_norm": 0.15541157126426697, |
|
"learning_rate": 9.340659340659341e-06, |
|
"loss": 1.4269, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.06750392464678179, |
|
"grad_norm": 0.1717188060283661, |
|
"learning_rate": 9.33281004709576e-06, |
|
"loss": 1.4827, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.06828885400313972, |
|
"grad_norm": 0.15374824404716492, |
|
"learning_rate": 9.324960753532182e-06, |
|
"loss": 1.4348, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.06907378335949764, |
|
"grad_norm": 0.16950677335262299, |
|
"learning_rate": 9.317111459968604e-06, |
|
"loss": 1.4779, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.06985871271585557, |
|
"grad_norm": 0.16479431092739105, |
|
"learning_rate": 9.309262166405025e-06, |
|
"loss": 1.4517, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.0706436420722135, |
|
"grad_norm": 0.16450008749961853, |
|
"learning_rate": 9.301412872841445e-06, |
|
"loss": 1.4352, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.07142857142857142, |
|
"grad_norm": 0.1584760844707489, |
|
"learning_rate": 9.293563579277866e-06, |
|
"loss": 1.3687, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.07221350078492936, |
|
"grad_norm": 0.17088983952999115, |
|
"learning_rate": 9.285714285714288e-06, |
|
"loss": 1.403, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.07299843014128729, |
|
"grad_norm": 0.16017332673072815, |
|
"learning_rate": 9.277864992150707e-06, |
|
"loss": 1.3504, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.07378335949764521, |
|
"grad_norm": 0.22771817445755005, |
|
"learning_rate": 9.270015698587129e-06, |
|
"loss": 1.4361, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.07456828885400314, |
|
"grad_norm": 0.1651688814163208, |
|
"learning_rate": 9.262166405023548e-06, |
|
"loss": 1.3933, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.07535321821036106, |
|
"grad_norm": 0.185661181807518, |
|
"learning_rate": 9.25431711145997e-06, |
|
"loss": 1.4446, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.076138147566719, |
|
"grad_norm": 0.17134390771389008, |
|
"learning_rate": 9.24646781789639e-06, |
|
"loss": 1.3812, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.07692307692307693, |
|
"grad_norm": 0.17040328681468964, |
|
"learning_rate": 9.238618524332811e-06, |
|
"loss": 1.393, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.07770800627943485, |
|
"grad_norm": 0.16197283565998077, |
|
"learning_rate": 9.230769230769232e-06, |
|
"loss": 1.3515, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.07849293563579278, |
|
"grad_norm": 0.18512940406799316, |
|
"learning_rate": 9.222919937205652e-06, |
|
"loss": 1.3884, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.07927786499215071, |
|
"grad_norm": 0.2023470401763916, |
|
"learning_rate": 9.215070643642073e-06, |
|
"loss": 1.4122, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.08006279434850863, |
|
"grad_norm": 0.1792641282081604, |
|
"learning_rate": 9.207221350078493e-06, |
|
"loss": 1.3936, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.08084772370486656, |
|
"grad_norm": 0.18946573138237, |
|
"learning_rate": 9.199372056514915e-06, |
|
"loss": 1.3314, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.08163265306122448, |
|
"grad_norm": 0.17217950522899628, |
|
"learning_rate": 9.191522762951334e-06, |
|
"loss": 1.3624, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.08241758241758242, |
|
"grad_norm": 0.1832340806722641, |
|
"learning_rate": 9.183673469387756e-06, |
|
"loss": 1.4058, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.08320251177394035, |
|
"grad_norm": 0.1757064312696457, |
|
"learning_rate": 9.175824175824175e-06, |
|
"loss": 1.3566, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.08398744113029827, |
|
"grad_norm": 0.17505265772342682, |
|
"learning_rate": 9.167974882260597e-06, |
|
"loss": 1.3217, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.0847723704866562, |
|
"grad_norm": 0.1719823032617569, |
|
"learning_rate": 9.160125588697018e-06, |
|
"loss": 1.3686, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.08555729984301413, |
|
"grad_norm": 0.16589918732643127, |
|
"learning_rate": 9.15227629513344e-06, |
|
"loss": 1.3998, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.08634222919937205, |
|
"grad_norm": 0.18314798176288605, |
|
"learning_rate": 9.14442700156986e-06, |
|
"loss": 1.399, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.08712715855572999, |
|
"grad_norm": 0.18713513016700745, |
|
"learning_rate": 9.13657770800628e-06, |
|
"loss": 1.4139, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.08791208791208792, |
|
"grad_norm": 0.16822576522827148, |
|
"learning_rate": 9.128728414442702e-06, |
|
"loss": 1.3247, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.08869701726844584, |
|
"grad_norm": 0.16645370423793793, |
|
"learning_rate": 9.120879120879122e-06, |
|
"loss": 1.3382, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.08948194662480377, |
|
"grad_norm": 0.16858340799808502, |
|
"learning_rate": 9.113029827315543e-06, |
|
"loss": 1.366, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.09026687598116169, |
|
"grad_norm": 0.15871913731098175, |
|
"learning_rate": 9.105180533751963e-06, |
|
"loss": 1.3446, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.09105180533751962, |
|
"grad_norm": 0.17642484605312347, |
|
"learning_rate": 9.097331240188384e-06, |
|
"loss": 1.4424, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.09183673469387756, |
|
"grad_norm": 0.16072145104408264, |
|
"learning_rate": 9.089481946624804e-06, |
|
"loss": 1.3247, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.09262166405023547, |
|
"grad_norm": 0.1545998454093933, |
|
"learning_rate": 9.081632653061225e-06, |
|
"loss": 1.3285, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.09340659340659341, |
|
"grad_norm": 0.15946722030639648, |
|
"learning_rate": 9.073783359497645e-06, |
|
"loss": 1.3123, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.09419152276295134, |
|
"grad_norm": 0.16009531915187836, |
|
"learning_rate": 9.065934065934067e-06, |
|
"loss": 1.3456, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.09497645211930926, |
|
"grad_norm": 0.15958304703235626, |
|
"learning_rate": 9.058084772370488e-06, |
|
"loss": 1.3039, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.09576138147566719, |
|
"grad_norm": 0.14737734198570251, |
|
"learning_rate": 9.050235478806908e-06, |
|
"loss": 1.3443, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.09654631083202511, |
|
"grad_norm": 0.15155759453773499, |
|
"learning_rate": 9.042386185243329e-06, |
|
"loss": 1.3459, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.09733124018838304, |
|
"grad_norm": 0.15477906167507172, |
|
"learning_rate": 9.034536891679749e-06, |
|
"loss": 1.3314, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.09811616954474098, |
|
"grad_norm": 0.14952191710472107, |
|
"learning_rate": 9.02668759811617e-06, |
|
"loss": 1.2907, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.0989010989010989, |
|
"grad_norm": 0.14685417711734772, |
|
"learning_rate": 9.01883830455259e-06, |
|
"loss": 1.3356, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.09968602825745683, |
|
"grad_norm": 0.1500052958726883, |
|
"learning_rate": 9.010989010989011e-06, |
|
"loss": 1.2724, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.10047095761381476, |
|
"grad_norm": 0.1418740600347519, |
|
"learning_rate": 9.003139717425433e-06, |
|
"loss": 1.3153, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.10125588697017268, |
|
"grad_norm": 0.15048982203006744, |
|
"learning_rate": 8.995290423861854e-06, |
|
"loss": 1.317, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.10204081632653061, |
|
"grad_norm": 0.1451820433139801, |
|
"learning_rate": 8.987441130298274e-06, |
|
"loss": 1.2907, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.10282574568288853, |
|
"grad_norm": 0.1553954780101776, |
|
"learning_rate": 8.979591836734695e-06, |
|
"loss": 1.3496, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.10361067503924647, |
|
"grad_norm": 0.13944025337696075, |
|
"learning_rate": 8.971742543171117e-06, |
|
"loss": 1.2967, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.1043956043956044, |
|
"grad_norm": 0.1505189687013626, |
|
"learning_rate": 8.963893249607536e-06, |
|
"loss": 1.242, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.10518053375196232, |
|
"grad_norm": 0.17538291215896606, |
|
"learning_rate": 8.956043956043958e-06, |
|
"loss": 1.3652, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.10596546310832025, |
|
"grad_norm": 0.1521396040916443, |
|
"learning_rate": 8.948194662480377e-06, |
|
"loss": 1.303, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.10675039246467818, |
|
"grad_norm": 0.14989960193634033, |
|
"learning_rate": 8.940345368916799e-06, |
|
"loss": 1.3348, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.1075353218210361, |
|
"grad_norm": 0.14487774670124054, |
|
"learning_rate": 8.932496075353219e-06, |
|
"loss": 1.307, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.10832025117739404, |
|
"grad_norm": 0.1494406759738922, |
|
"learning_rate": 8.92464678178964e-06, |
|
"loss": 1.3126, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.10910518053375197, |
|
"grad_norm": 0.16444607079029083, |
|
"learning_rate": 8.91679748822606e-06, |
|
"loss": 1.2694, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.10989010989010989, |
|
"grad_norm": 0.15526680648326874, |
|
"learning_rate": 8.908948194662481e-06, |
|
"loss": 1.2646, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.11067503924646782, |
|
"grad_norm": 0.15668538212776184, |
|
"learning_rate": 8.9010989010989e-06, |
|
"loss": 1.2587, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.11145996860282574, |
|
"grad_norm": 0.14996470510959625, |
|
"learning_rate": 8.893249607535322e-06, |
|
"loss": 1.3122, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.11224489795918367, |
|
"grad_norm": 0.14894387125968933, |
|
"learning_rate": 8.885400313971744e-06, |
|
"loss": 1.2909, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.1130298273155416, |
|
"grad_norm": 0.18087385594844818, |
|
"learning_rate": 8.877551020408163e-06, |
|
"loss": 1.3161, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.11381475667189953, |
|
"grad_norm": 0.15965577960014343, |
|
"learning_rate": 8.869701726844585e-06, |
|
"loss": 1.2683, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.11459968602825746, |
|
"grad_norm": 0.15005330741405487, |
|
"learning_rate": 8.861852433281004e-06, |
|
"loss": 1.3083, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.11538461538461539, |
|
"grad_norm": 0.14711208641529083, |
|
"learning_rate": 8.854003139717426e-06, |
|
"loss": 1.3019, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.11616954474097331, |
|
"grad_norm": 0.14721107482910156, |
|
"learning_rate": 8.846153846153847e-06, |
|
"loss": 1.2469, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.11695447409733124, |
|
"grad_norm": 0.16075782477855682, |
|
"learning_rate": 8.838304552590269e-06, |
|
"loss": 1.2696, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.11773940345368916, |
|
"grad_norm": 0.14981156587600708, |
|
"learning_rate": 8.830455259026688e-06, |
|
"loss": 1.2416, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1185243328100471, |
|
"grad_norm": 0.15048524737358093, |
|
"learning_rate": 8.82260596546311e-06, |
|
"loss": 1.235, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.11930926216640503, |
|
"grad_norm": 0.1528954803943634, |
|
"learning_rate": 8.81475667189953e-06, |
|
"loss": 1.3038, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.12009419152276295, |
|
"grad_norm": 0.15498745441436768, |
|
"learning_rate": 8.80690737833595e-06, |
|
"loss": 1.3014, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.12087912087912088, |
|
"grad_norm": 0.15970808267593384, |
|
"learning_rate": 8.799058084772372e-06, |
|
"loss": 1.2824, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.12166405023547881, |
|
"grad_norm": 0.15735112130641937, |
|
"learning_rate": 8.791208791208792e-06, |
|
"loss": 1.2684, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.12244897959183673, |
|
"grad_norm": 0.15894393622875214, |
|
"learning_rate": 8.783359497645213e-06, |
|
"loss": 1.2283, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.12323390894819466, |
|
"grad_norm": 0.16271716356277466, |
|
"learning_rate": 8.775510204081633e-06, |
|
"loss": 1.2974, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.12401883830455258, |
|
"grad_norm": 0.15001676976680756, |
|
"learning_rate": 8.767660910518054e-06, |
|
"loss": 1.2688, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.12480376766091052, |
|
"grad_norm": 0.1491970270872116, |
|
"learning_rate": 8.759811616954474e-06, |
|
"loss": 1.2482, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.12558869701726844, |
|
"grad_norm": 0.15794126689434052, |
|
"learning_rate": 8.751962323390896e-06, |
|
"loss": 1.2605, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.12637362637362637, |
|
"grad_norm": 0.14844392240047455, |
|
"learning_rate": 8.744113029827315e-06, |
|
"loss": 1.2332, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.1271585557299843, |
|
"grad_norm": 0.16946828365325928, |
|
"learning_rate": 8.736263736263737e-06, |
|
"loss": 1.3136, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.12794348508634223, |
|
"grad_norm": 0.14547237753868103, |
|
"learning_rate": 8.728414442700158e-06, |
|
"loss": 1.2409, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.12872841444270017, |
|
"grad_norm": 0.15002243220806122, |
|
"learning_rate": 8.720565149136578e-06, |
|
"loss": 1.1916, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.12951334379905807, |
|
"grad_norm": 0.1735372096300125, |
|
"learning_rate": 8.712715855573e-06, |
|
"loss": 1.2371, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.130298273155416, |
|
"grad_norm": 0.18110059201717377, |
|
"learning_rate": 8.704866562009419e-06, |
|
"loss": 1.3246, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.13108320251177394, |
|
"grad_norm": 0.16027387976646423, |
|
"learning_rate": 8.69701726844584e-06, |
|
"loss": 1.2489, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.13186813186813187, |
|
"grad_norm": 0.15014226734638214, |
|
"learning_rate": 8.689167974882262e-06, |
|
"loss": 1.2068, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.1326530612244898, |
|
"grad_norm": 0.15863986313343048, |
|
"learning_rate": 8.681318681318681e-06, |
|
"loss": 1.2825, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.13343799058084774, |
|
"grad_norm": 0.17867471277713776, |
|
"learning_rate": 8.673469387755103e-06, |
|
"loss": 1.2607, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.13422291993720564, |
|
"grad_norm": 0.1596807986497879, |
|
"learning_rate": 8.665620094191524e-06, |
|
"loss": 1.2067, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.13500784929356358, |
|
"grad_norm": 0.1603277176618576, |
|
"learning_rate": 8.657770800627944e-06, |
|
"loss": 1.2414, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.1357927786499215, |
|
"grad_norm": 0.26285287737846375, |
|
"learning_rate": 8.649921507064365e-06, |
|
"loss": 1.1949, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.13657770800627944, |
|
"grad_norm": 0.15682700276374817, |
|
"learning_rate": 8.642072213500787e-06, |
|
"loss": 1.1832, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.13736263736263737, |
|
"grad_norm": 0.16141754388809204, |
|
"learning_rate": 8.634222919937206e-06, |
|
"loss": 1.2334, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.13814756671899528, |
|
"grad_norm": 0.18615436553955078, |
|
"learning_rate": 8.626373626373628e-06, |
|
"loss": 1.256, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.1389324960753532, |
|
"grad_norm": 0.15746115148067474, |
|
"learning_rate": 8.618524332810048e-06, |
|
"loss": 1.2267, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.13971742543171115, |
|
"grad_norm": 0.16463595628738403, |
|
"learning_rate": 8.610675039246469e-06, |
|
"loss": 1.2638, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.14050235478806908, |
|
"grad_norm": 0.15357612073421478, |
|
"learning_rate": 8.602825745682889e-06, |
|
"loss": 1.2262, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.141287284144427, |
|
"grad_norm": 0.16407904028892517, |
|
"learning_rate": 8.59497645211931e-06, |
|
"loss": 1.1631, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.14207221350078492, |
|
"grad_norm": 0.15864725410938263, |
|
"learning_rate": 8.58712715855573e-06, |
|
"loss": 1.1992, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.14285714285714285, |
|
"grad_norm": 0.16168276965618134, |
|
"learning_rate": 8.579277864992151e-06, |
|
"loss": 1.2385, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.14364207221350078, |
|
"grad_norm": 0.15585263073444366, |
|
"learning_rate": 8.571428571428571e-06, |
|
"loss": 1.2105, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.14442700156985872, |
|
"grad_norm": 0.15923231840133667, |
|
"learning_rate": 8.563579277864992e-06, |
|
"loss": 1.2208, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.14521193092621665, |
|
"grad_norm": 0.1669979989528656, |
|
"learning_rate": 8.555729984301414e-06, |
|
"loss": 1.2528, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.14599686028257458, |
|
"grad_norm": 0.1686784327030182, |
|
"learning_rate": 8.547880690737833e-06, |
|
"loss": 1.2136, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.14678178963893249, |
|
"grad_norm": 0.16007350385189056, |
|
"learning_rate": 8.540031397174255e-06, |
|
"loss": 1.2197, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.14756671899529042, |
|
"grad_norm": 0.15759383141994476, |
|
"learning_rate": 8.532182103610676e-06, |
|
"loss": 1.2286, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.14835164835164835, |
|
"grad_norm": 0.1621437519788742, |
|
"learning_rate": 8.524332810047096e-06, |
|
"loss": 1.1928, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.14913657770800628, |
|
"grad_norm": 0.1673436015844345, |
|
"learning_rate": 8.516483516483517e-06, |
|
"loss": 1.2709, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.14992150706436422, |
|
"grad_norm": 0.15842801332473755, |
|
"learning_rate": 8.508634222919939e-06, |
|
"loss": 1.213, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.15070643642072212, |
|
"grad_norm": 0.16606052219867706, |
|
"learning_rate": 8.500784929356358e-06, |
|
"loss": 1.2274, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.15149136577708006, |
|
"grad_norm": 0.24337609112262726, |
|
"learning_rate": 8.49293563579278e-06, |
|
"loss": 1.2763, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.152276295133438, |
|
"grad_norm": 0.1524026244878769, |
|
"learning_rate": 8.4850863422292e-06, |
|
"loss": 1.1981, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.15306122448979592, |
|
"grad_norm": 0.16204509139060974, |
|
"learning_rate": 8.477237048665621e-06, |
|
"loss": 1.2361, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.15384615384615385, |
|
"grad_norm": 0.16638584434986115, |
|
"learning_rate": 8.469387755102042e-06, |
|
"loss": 1.2309, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.1546310832025118, |
|
"grad_norm": 0.16714318096637726, |
|
"learning_rate": 8.461538461538462e-06, |
|
"loss": 1.2692, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.1554160125588697, |
|
"grad_norm": 0.168153315782547, |
|
"learning_rate": 8.453689167974884e-06, |
|
"loss": 1.2476, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.15620094191522763, |
|
"grad_norm": 0.1681162267923355, |
|
"learning_rate": 8.445839874411303e-06, |
|
"loss": 1.2504, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.15698587127158556, |
|
"grad_norm": 0.16602487862110138, |
|
"learning_rate": 8.437990580847725e-06, |
|
"loss": 1.2636, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.1577708006279435, |
|
"grad_norm": 0.1714046150445938, |
|
"learning_rate": 8.430141287284144e-06, |
|
"loss": 1.2159, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.15855572998430142, |
|
"grad_norm": 0.168313130736351, |
|
"learning_rate": 8.422291993720566e-06, |
|
"loss": 1.1932, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.15934065934065933, |
|
"grad_norm": 0.1813940554857254, |
|
"learning_rate": 8.414442700156985e-06, |
|
"loss": 1.2291, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.16012558869701726, |
|
"grad_norm": 0.16745160520076752, |
|
"learning_rate": 8.406593406593407e-06, |
|
"loss": 1.2088, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.1609105180533752, |
|
"grad_norm": 0.1752898395061493, |
|
"learning_rate": 8.398744113029828e-06, |
|
"loss": 1.2324, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.16169544740973313, |
|
"grad_norm": 0.1678934097290039, |
|
"learning_rate": 8.390894819466248e-06, |
|
"loss": 1.2545, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.16248037676609106, |
|
"grad_norm": 0.1756390780210495, |
|
"learning_rate": 8.38304552590267e-06, |
|
"loss": 1.2421, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.16326530612244897, |
|
"grad_norm": 0.17289415001869202, |
|
"learning_rate": 8.37519623233909e-06, |
|
"loss": 1.2063, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.1640502354788069, |
|
"grad_norm": 0.17906337976455688, |
|
"learning_rate": 8.36734693877551e-06, |
|
"loss": 1.2598, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.16483516483516483, |
|
"grad_norm": 0.17102067172527313, |
|
"learning_rate": 8.359497645211932e-06, |
|
"loss": 1.1696, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.16562009419152277, |
|
"grad_norm": 0.16523365676403046, |
|
"learning_rate": 8.351648351648353e-06, |
|
"loss": 1.1769, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.1664050235478807, |
|
"grad_norm": 0.1749919205904007, |
|
"learning_rate": 8.343799058084773e-06, |
|
"loss": 1.1784, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.16718995290423863, |
|
"grad_norm": 0.17963068187236786, |
|
"learning_rate": 8.335949764521194e-06, |
|
"loss": 1.1859, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.16797488226059654, |
|
"grad_norm": 0.17333124577999115, |
|
"learning_rate": 8.328100470957614e-06, |
|
"loss": 1.2148, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.16875981161695447, |
|
"grad_norm": 0.17331229150295258, |
|
"learning_rate": 8.320251177394036e-06, |
|
"loss": 1.2404, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.1695447409733124, |
|
"grad_norm": 0.23610031604766846, |
|
"learning_rate": 8.312401883830455e-06, |
|
"loss": 1.222, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.17032967032967034, |
|
"grad_norm": 0.19044527411460876, |
|
"learning_rate": 8.304552590266877e-06, |
|
"loss": 1.1923, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.17111459968602827, |
|
"grad_norm": 0.16924557089805603, |
|
"learning_rate": 8.296703296703298e-06, |
|
"loss": 1.2481, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.17189952904238617, |
|
"grad_norm": 0.17461296916007996, |
|
"learning_rate": 8.288854003139718e-06, |
|
"loss": 1.204, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.1726844583987441, |
|
"grad_norm": 0.16900260746479034, |
|
"learning_rate": 8.281004709576139e-06, |
|
"loss": 1.158, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.17346938775510204, |
|
"grad_norm": 0.1738884598016739, |
|
"learning_rate": 8.273155416012559e-06, |
|
"loss": 1.1835, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.17425431711145997, |
|
"grad_norm": 0.18512043356895447, |
|
"learning_rate": 8.26530612244898e-06, |
|
"loss": 1.202, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.1750392464678179, |
|
"grad_norm": 0.15882278978824615, |
|
"learning_rate": 8.2574568288854e-06, |
|
"loss": 1.1717, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.17582417582417584, |
|
"grad_norm": 0.1748121678829193, |
|
"learning_rate": 8.249607535321821e-06, |
|
"loss": 1.2407, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.17660910518053374, |
|
"grad_norm": 0.17194059491157532, |
|
"learning_rate": 8.241758241758243e-06, |
|
"loss": 1.2104, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.17739403453689168, |
|
"grad_norm": 0.17927075922489166, |
|
"learning_rate": 8.233908948194662e-06, |
|
"loss": 1.1295, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.1781789638932496, |
|
"grad_norm": 0.17285114526748657, |
|
"learning_rate": 8.226059654631084e-06, |
|
"loss": 1.1364, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.17896389324960754, |
|
"grad_norm": 0.18146753311157227, |
|
"learning_rate": 8.218210361067505e-06, |
|
"loss": 1.2123, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.17974882260596547, |
|
"grad_norm": 0.17958636581897736, |
|
"learning_rate": 8.210361067503925e-06, |
|
"loss": 1.2155, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.18053375196232338, |
|
"grad_norm": 0.1809559315443039, |
|
"learning_rate": 8.202511773940346e-06, |
|
"loss": 1.2193, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.1813186813186813, |
|
"grad_norm": 0.17428240180015564, |
|
"learning_rate": 8.194662480376768e-06, |
|
"loss": 1.1756, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.18210361067503925, |
|
"grad_norm": 0.16993722319602966, |
|
"learning_rate": 8.186813186813188e-06, |
|
"loss": 1.1373, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.18288854003139718, |
|
"grad_norm": 0.16454678773880005, |
|
"learning_rate": 8.178963893249609e-06, |
|
"loss": 1.1778, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.1836734693877551, |
|
"grad_norm": 0.19709423184394836, |
|
"learning_rate": 8.171114599686029e-06, |
|
"loss": 1.1672, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.18445839874411302, |
|
"grad_norm": 0.17396849393844604, |
|
"learning_rate": 8.16326530612245e-06, |
|
"loss": 1.1614, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.18524332810047095, |
|
"grad_norm": 0.17194689810276031, |
|
"learning_rate": 8.15541601255887e-06, |
|
"loss": 1.1568, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.18602825745682888, |
|
"grad_norm": 0.17015020549297333, |
|
"learning_rate": 8.147566718995291e-06, |
|
"loss": 1.1737, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.18681318681318682, |
|
"grad_norm": 0.181587353348732, |
|
"learning_rate": 8.139717425431711e-06, |
|
"loss": 1.1874, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.18759811616954475, |
|
"grad_norm": 0.17715583741664886, |
|
"learning_rate": 8.131868131868132e-06, |
|
"loss": 1.2042, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.18838304552590268, |
|
"grad_norm": 0.18029291927814484, |
|
"learning_rate": 8.124018838304554e-06, |
|
"loss": 1.2087, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.1891679748822606, |
|
"grad_norm": 0.1827882081270218, |
|
"learning_rate": 8.116169544740973e-06, |
|
"loss": 1.1878, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.18995290423861852, |
|
"grad_norm": 0.18994055688381195, |
|
"learning_rate": 8.108320251177395e-06, |
|
"loss": 1.2756, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.19073783359497645, |
|
"grad_norm": 0.18137842416763306, |
|
"learning_rate": 8.100470957613814e-06, |
|
"loss": 1.1633, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.19152276295133439, |
|
"grad_norm": 0.18730993568897247, |
|
"learning_rate": 8.092621664050236e-06, |
|
"loss": 1.1659, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.19230769230769232, |
|
"grad_norm": 0.17532621324062347, |
|
"learning_rate": 8.084772370486657e-06, |
|
"loss": 1.212, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.19309262166405022, |
|
"grad_norm": 0.18628445267677307, |
|
"learning_rate": 8.076923076923077e-06, |
|
"loss": 1.1684, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.19387755102040816, |
|
"grad_norm": 0.1802094727754593, |
|
"learning_rate": 8.069073783359498e-06, |
|
"loss": 1.2242, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.1946624803767661, |
|
"grad_norm": 0.18173186480998993, |
|
"learning_rate": 8.06122448979592e-06, |
|
"loss": 1.2428, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.19544740973312402, |
|
"grad_norm": 0.17413657903671265, |
|
"learning_rate": 8.05337519623234e-06, |
|
"loss": 1.1303, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.19623233908948196, |
|
"grad_norm": 0.18897269666194916, |
|
"learning_rate": 8.045525902668761e-06, |
|
"loss": 1.2011, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.1970172684458399, |
|
"grad_norm": 0.18273556232452393, |
|
"learning_rate": 8.037676609105182e-06, |
|
"loss": 1.195, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.1978021978021978, |
|
"grad_norm": 0.17363931238651276, |
|
"learning_rate": 8.029827315541602e-06, |
|
"loss": 1.151, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.19858712715855573, |
|
"grad_norm": 0.19355787336826324, |
|
"learning_rate": 8.021978021978023e-06, |
|
"loss": 1.1883, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.19937205651491366, |
|
"grad_norm": 0.1852579414844513, |
|
"learning_rate": 8.014128728414443e-06, |
|
"loss": 1.1829, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.2001569858712716, |
|
"grad_norm": 0.18897081911563873, |
|
"learning_rate": 8.006279434850865e-06, |
|
"loss": 1.1633, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.20094191522762953, |
|
"grad_norm": 0.177345871925354, |
|
"learning_rate": 7.998430141287284e-06, |
|
"loss": 1.1543, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.20172684458398743, |
|
"grad_norm": 0.19197995960712433, |
|
"learning_rate": 7.990580847723706e-06, |
|
"loss": 1.1124, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.20251177394034536, |
|
"grad_norm": 0.1918957382440567, |
|
"learning_rate": 7.982731554160125e-06, |
|
"loss": 1.2175, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.2032967032967033, |
|
"grad_norm": 0.3751870095729828, |
|
"learning_rate": 7.974882260596547e-06, |
|
"loss": 1.208, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.20408163265306123, |
|
"grad_norm": 0.19366663694381714, |
|
"learning_rate": 7.967032967032966e-06, |
|
"loss": 1.1758, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.20486656200941916, |
|
"grad_norm": 0.1843341588973999, |
|
"learning_rate": 7.959183673469388e-06, |
|
"loss": 1.1385, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.20565149136577707, |
|
"grad_norm": 0.17850859463214874, |
|
"learning_rate": 7.95133437990581e-06, |
|
"loss": 1.1707, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.206436420722135, |
|
"grad_norm": 0.21179649233818054, |
|
"learning_rate": 7.943485086342229e-06, |
|
"loss": 1.2079, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.20722135007849293, |
|
"grad_norm": 0.21911092102527618, |
|
"learning_rate": 7.93563579277865e-06, |
|
"loss": 1.2071, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.20800627943485087, |
|
"grad_norm": 0.1800030916929245, |
|
"learning_rate": 7.927786499215072e-06, |
|
"loss": 1.1897, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.2087912087912088, |
|
"grad_norm": 0.20178881287574768, |
|
"learning_rate": 7.919937205651492e-06, |
|
"loss": 1.1307, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.20957613814756673, |
|
"grad_norm": 0.18904000520706177, |
|
"learning_rate": 7.912087912087913e-06, |
|
"loss": 1.1234, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.21036106750392464, |
|
"grad_norm": 0.18753568828105927, |
|
"learning_rate": 7.904238618524334e-06, |
|
"loss": 1.1312, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.21114599686028257, |
|
"grad_norm": 0.18907895684242249, |
|
"learning_rate": 7.896389324960754e-06, |
|
"loss": 1.2075, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.2119309262166405, |
|
"grad_norm": 0.20011885464191437, |
|
"learning_rate": 7.888540031397175e-06, |
|
"loss": 1.1175, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.21271585557299844, |
|
"grad_norm": 0.1855919063091278, |
|
"learning_rate": 7.880690737833597e-06, |
|
"loss": 1.1581, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.21350078492935637, |
|
"grad_norm": 0.18904899060726166, |
|
"learning_rate": 7.872841444270017e-06, |
|
"loss": 1.1513, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.21428571428571427, |
|
"grad_norm": 0.18176890909671783, |
|
"learning_rate": 7.864992150706438e-06, |
|
"loss": 1.1448, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.2150706436420722, |
|
"grad_norm": 0.18389371037483215, |
|
"learning_rate": 7.857142857142858e-06, |
|
"loss": 1.1695, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.21585557299843014, |
|
"grad_norm": 0.1845601201057434, |
|
"learning_rate": 7.849293563579279e-06, |
|
"loss": 1.1295, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.21664050235478807, |
|
"grad_norm": 0.2147328108549118, |
|
"learning_rate": 7.841444270015699e-06, |
|
"loss": 1.1369, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.217425431711146, |
|
"grad_norm": 0.20094560086727142, |
|
"learning_rate": 7.83359497645212e-06, |
|
"loss": 1.1666, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.21821036106750394, |
|
"grad_norm": 0.1994454562664032, |
|
"learning_rate": 7.82574568288854e-06, |
|
"loss": 1.171, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.21899529042386184, |
|
"grad_norm": 0.19250410795211792, |
|
"learning_rate": 7.817896389324961e-06, |
|
"loss": 1.1709, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.21978021978021978, |
|
"grad_norm": 0.20511236786842346, |
|
"learning_rate": 7.810047095761381e-06, |
|
"loss": 1.1788, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.2205651491365777, |
|
"grad_norm": 0.2156459391117096, |
|
"learning_rate": 7.802197802197802e-06, |
|
"loss": 1.1543, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.22135007849293564, |
|
"grad_norm": 0.18635962903499603, |
|
"learning_rate": 7.794348508634224e-06, |
|
"loss": 1.1375, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.22213500784929358, |
|
"grad_norm": 0.1912240833044052, |
|
"learning_rate": 7.786499215070644e-06, |
|
"loss": 1.1901, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.22291993720565148, |
|
"grad_norm": 0.19378046691417694, |
|
"learning_rate": 7.778649921507065e-06, |
|
"loss": 1.2012, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.2237048665620094, |
|
"grad_norm": 0.19641022384166718, |
|
"learning_rate": 7.770800627943486e-06, |
|
"loss": 1.1817, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.22448979591836735, |
|
"grad_norm": 0.1922275871038437, |
|
"learning_rate": 7.762951334379906e-06, |
|
"loss": 1.182, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.22527472527472528, |
|
"grad_norm": 0.19869813323020935, |
|
"learning_rate": 7.755102040816327e-06, |
|
"loss": 1.1975, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.2260596546310832, |
|
"grad_norm": 0.183540940284729, |
|
"learning_rate": 7.747252747252749e-06, |
|
"loss": 1.2028, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.22684458398744112, |
|
"grad_norm": 0.20018337666988373, |
|
"learning_rate": 7.739403453689169e-06, |
|
"loss": 1.2147, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.22762951334379905, |
|
"grad_norm": 0.2085750699043274, |
|
"learning_rate": 7.73155416012559e-06, |
|
"loss": 1.1198, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.22841444270015698, |
|
"grad_norm": 0.19977134466171265, |
|
"learning_rate": 7.72370486656201e-06, |
|
"loss": 1.1528, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.22919937205651492, |
|
"grad_norm": 0.19265629351139069, |
|
"learning_rate": 7.715855572998431e-06, |
|
"loss": 1.1401, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.22998430141287285, |
|
"grad_norm": 0.191536083817482, |
|
"learning_rate": 7.708006279434852e-06, |
|
"loss": 1.1556, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.23076923076923078, |
|
"grad_norm": 0.2022550255060196, |
|
"learning_rate": 7.700156985871272e-06, |
|
"loss": 1.0994, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.2315541601255887, |
|
"grad_norm": 0.1965666264295578, |
|
"learning_rate": 7.692307692307694e-06, |
|
"loss": 1.0851, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.23233908948194662, |
|
"grad_norm": 0.23619407415390015, |
|
"learning_rate": 7.684458398744113e-06, |
|
"loss": 1.2371, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.23312401883830455, |
|
"grad_norm": 0.18914885818958282, |
|
"learning_rate": 7.676609105180535e-06, |
|
"loss": 1.1038, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.23390894819466249, |
|
"grad_norm": 0.18863645195960999, |
|
"learning_rate": 7.668759811616954e-06, |
|
"loss": 1.1318, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.23469387755102042, |
|
"grad_norm": 0.2066003382205963, |
|
"learning_rate": 7.660910518053376e-06, |
|
"loss": 1.1542, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.23547880690737832, |
|
"grad_norm": 0.21004410088062286, |
|
"learning_rate": 7.653061224489796e-06, |
|
"loss": 1.1647, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.23626373626373626, |
|
"grad_norm": 0.19493690133094788, |
|
"learning_rate": 7.645211930926217e-06, |
|
"loss": 1.1743, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.2370486656200942, |
|
"grad_norm": 0.1926044225692749, |
|
"learning_rate": 7.637362637362638e-06, |
|
"loss": 1.1333, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.23783359497645212, |
|
"grad_norm": 0.19471141695976257, |
|
"learning_rate": 7.629513343799058e-06, |
|
"loss": 1.0989, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.23861852433281006, |
|
"grad_norm": 0.1914125680923462, |
|
"learning_rate": 7.62166405023548e-06, |
|
"loss": 1.1427, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.239403453689168, |
|
"grad_norm": 0.21529747545719147, |
|
"learning_rate": 7.6138147566719e-06, |
|
"loss": 1.117, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.2401883830455259, |
|
"grad_norm": 0.2219187617301941, |
|
"learning_rate": 7.605965463108321e-06, |
|
"loss": 1.1584, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.24097331240188383, |
|
"grad_norm": 0.20323887467384338, |
|
"learning_rate": 7.598116169544741e-06, |
|
"loss": 1.1543, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.24175824175824176, |
|
"grad_norm": 0.19014237821102142, |
|
"learning_rate": 7.5902668759811625e-06, |
|
"loss": 1.1149, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.2425431711145997, |
|
"grad_norm": 0.2022504359483719, |
|
"learning_rate": 7.582417582417583e-06, |
|
"loss": 1.1835, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.24332810047095763, |
|
"grad_norm": 0.19809505343437195, |
|
"learning_rate": 7.574568288854004e-06, |
|
"loss": 1.0916, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.24411302982731553, |
|
"grad_norm": 0.19800910353660583, |
|
"learning_rate": 7.566718995290424e-06, |
|
"loss": 1.1467, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.24489795918367346, |
|
"grad_norm": 0.19982385635375977, |
|
"learning_rate": 7.558869701726846e-06, |
|
"loss": 1.1186, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.2456828885400314, |
|
"grad_norm": 0.19915273785591125, |
|
"learning_rate": 7.551020408163265e-06, |
|
"loss": 1.1743, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.24646781789638933, |
|
"grad_norm": 0.21240346133708954, |
|
"learning_rate": 7.543171114599687e-06, |
|
"loss": 1.1575, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.24725274725274726, |
|
"grad_norm": 0.19470131397247314, |
|
"learning_rate": 7.535321821036108e-06, |
|
"loss": 1.1745, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.24803767660910517, |
|
"grad_norm": 0.21574276685714722, |
|
"learning_rate": 7.527472527472528e-06, |
|
"loss": 1.1922, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.2488226059654631, |
|
"grad_norm": 0.1990724802017212, |
|
"learning_rate": 7.519623233908949e-06, |
|
"loss": 1.1765, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.24960753532182103, |
|
"grad_norm": 0.1972692310810089, |
|
"learning_rate": 7.511773940345369e-06, |
|
"loss": 1.1848, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.25039246467817894, |
|
"grad_norm": 0.21943879127502441, |
|
"learning_rate": 7.50392464678179e-06, |
|
"loss": 1.1103, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.25117739403453687, |
|
"grad_norm": 0.20259033143520355, |
|
"learning_rate": 7.496075353218211e-06, |
|
"loss": 1.1587, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.2519623233908948, |
|
"grad_norm": 0.20974001288414001, |
|
"learning_rate": 7.488226059654632e-06, |
|
"loss": 1.1161, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.25274725274725274, |
|
"grad_norm": 0.1978650540113449, |
|
"learning_rate": 7.480376766091052e-06, |
|
"loss": 1.1743, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.25353218210361067, |
|
"grad_norm": 0.21677835285663605, |
|
"learning_rate": 7.472527472527473e-06, |
|
"loss": 1.1915, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.2543171114599686, |
|
"grad_norm": 0.19082266092300415, |
|
"learning_rate": 7.464678178963893e-06, |
|
"loss": 1.1146, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.25510204081632654, |
|
"grad_norm": 0.20703904330730438, |
|
"learning_rate": 7.4568288854003145e-06, |
|
"loss": 1.1325, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.25588697017268447, |
|
"grad_norm": 0.20791569352149963, |
|
"learning_rate": 7.448979591836736e-06, |
|
"loss": 1.1569, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.2566718995290424, |
|
"grad_norm": 0.2217319905757904, |
|
"learning_rate": 7.441130298273156e-06, |
|
"loss": 1.1823, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.25745682888540034, |
|
"grad_norm": 0.20722460746765137, |
|
"learning_rate": 7.433281004709577e-06, |
|
"loss": 1.1482, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.25824175824175827, |
|
"grad_norm": 0.2060956358909607, |
|
"learning_rate": 7.425431711145998e-06, |
|
"loss": 1.1829, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.25902668759811615, |
|
"grad_norm": 0.21304012835025787, |
|
"learning_rate": 7.417582417582418e-06, |
|
"loss": 1.1984, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.2598116169544741, |
|
"grad_norm": 0.21216687560081482, |
|
"learning_rate": 7.409733124018839e-06, |
|
"loss": 1.1238, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.260596546310832, |
|
"grad_norm": 0.2066863626241684, |
|
"learning_rate": 7.40188383045526e-06, |
|
"loss": 1.1304, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.26138147566718994, |
|
"grad_norm": 0.2035655975341797, |
|
"learning_rate": 7.39403453689168e-06, |
|
"loss": 1.1423, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.2621664050235479, |
|
"grad_norm": 0.2093246579170227, |
|
"learning_rate": 7.386185243328101e-06, |
|
"loss": 1.185, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.2629513343799058, |
|
"grad_norm": 0.20066991448402405, |
|
"learning_rate": 7.378335949764521e-06, |
|
"loss": 1.0917, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.26373626373626374, |
|
"grad_norm": 0.21041403710842133, |
|
"learning_rate": 7.370486656200942e-06, |
|
"loss": 1.1487, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.2645211930926217, |
|
"grad_norm": 0.19787679612636566, |
|
"learning_rate": 7.362637362637364e-06, |
|
"loss": 1.1429, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.2653061224489796, |
|
"grad_norm": 0.2156287282705307, |
|
"learning_rate": 7.3547880690737835e-06, |
|
"loss": 1.1162, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.26609105180533754, |
|
"grad_norm": 0.23158158361911774, |
|
"learning_rate": 7.346938775510205e-06, |
|
"loss": 1.1608, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.2668759811616955, |
|
"grad_norm": 0.21346524357795715, |
|
"learning_rate": 7.339089481946625e-06, |
|
"loss": 1.1092, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.26766091051805335, |
|
"grad_norm": 0.2354096919298172, |
|
"learning_rate": 7.331240188383047e-06, |
|
"loss": 1.1448, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.2684458398744113, |
|
"grad_norm": 0.20579902827739716, |
|
"learning_rate": 7.3233908948194665e-06, |
|
"loss": 1.1627, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.2692307692307692, |
|
"grad_norm": 0.24660111963748932, |
|
"learning_rate": 7.315541601255888e-06, |
|
"loss": 1.1113, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.27001569858712715, |
|
"grad_norm": 0.21383073925971985, |
|
"learning_rate": 7.307692307692308e-06, |
|
"loss": 1.1361, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.2708006279434851, |
|
"grad_norm": 0.2156330794095993, |
|
"learning_rate": 7.299843014128729e-06, |
|
"loss": 1.0608, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.271585557299843, |
|
"grad_norm": 0.20856386423110962, |
|
"learning_rate": 7.29199372056515e-06, |
|
"loss": 1.1545, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.27237048665620095, |
|
"grad_norm": 0.21835865080356598, |
|
"learning_rate": 7.28414442700157e-06, |
|
"loss": 1.1998, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.2731554160125589, |
|
"grad_norm": 0.23513004183769226, |
|
"learning_rate": 7.2762951334379916e-06, |
|
"loss": 1.2048, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.2739403453689168, |
|
"grad_norm": 0.22506913542747498, |
|
"learning_rate": 7.268445839874412e-06, |
|
"loss": 1.1828, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.27472527472527475, |
|
"grad_norm": 0.216568723320961, |
|
"learning_rate": 7.260596546310833e-06, |
|
"loss": 1.1928, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.2755102040816326, |
|
"grad_norm": 0.2345465123653412, |
|
"learning_rate": 7.252747252747253e-06, |
|
"loss": 1.1723, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.27629513343799056, |
|
"grad_norm": 0.21116310358047485, |
|
"learning_rate": 7.244897959183675e-06, |
|
"loss": 1.1362, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.2770800627943485, |
|
"grad_norm": 0.20539937913417816, |
|
"learning_rate": 7.237048665620094e-06, |
|
"loss": 1.1692, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.2778649921507064, |
|
"grad_norm": 0.22223587334156036, |
|
"learning_rate": 7.229199372056516e-06, |
|
"loss": 1.1589, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.27864992150706436, |
|
"grad_norm": 0.20060519874095917, |
|
"learning_rate": 7.2213500784929355e-06, |
|
"loss": 1.1264, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.2794348508634223, |
|
"grad_norm": 0.21847526729106903, |
|
"learning_rate": 7.213500784929357e-06, |
|
"loss": 1.1401, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.2802197802197802, |
|
"grad_norm": 0.22963590919971466, |
|
"learning_rate": 7.205651491365777e-06, |
|
"loss": 1.1523, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.28100470957613816, |
|
"grad_norm": 0.21631349623203278, |
|
"learning_rate": 7.197802197802198e-06, |
|
"loss": 1.1265, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.2817896389324961, |
|
"grad_norm": 0.21508990228176117, |
|
"learning_rate": 7.189952904238619e-06, |
|
"loss": 1.1125, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.282574568288854, |
|
"grad_norm": 0.2024298906326294, |
|
"learning_rate": 7.18210361067504e-06, |
|
"loss": 1.0864, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.28335949764521196, |
|
"grad_norm": 0.21588054299354553, |
|
"learning_rate": 7.174254317111461e-06, |
|
"loss": 1.1265, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.28414442700156983, |
|
"grad_norm": 0.3438448905944824, |
|
"learning_rate": 7.166405023547881e-06, |
|
"loss": 1.1087, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.28492935635792777, |
|
"grad_norm": 0.23037710785865784, |
|
"learning_rate": 7.1585557299843024e-06, |
|
"loss": 1.1683, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.2857142857142857, |
|
"grad_norm": 0.371964693069458, |
|
"learning_rate": 7.150706436420722e-06, |
|
"loss": 1.1306, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.28649921507064363, |
|
"grad_norm": 0.2277277410030365, |
|
"learning_rate": 7.1428571428571436e-06, |
|
"loss": 1.1758, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.28728414442700156, |
|
"grad_norm": 0.20543402433395386, |
|
"learning_rate": 7.135007849293564e-06, |
|
"loss": 1.1501, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.2880690737833595, |
|
"grad_norm": 0.2121914029121399, |
|
"learning_rate": 7.127158555729985e-06, |
|
"loss": 1.1247, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.28885400313971743, |
|
"grad_norm": 0.22225429117679596, |
|
"learning_rate": 7.119309262166405e-06, |
|
"loss": 1.1231, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.28963893249607536, |
|
"grad_norm": 0.21142037212848663, |
|
"learning_rate": 7.111459968602827e-06, |
|
"loss": 1.1528, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.2904238618524333, |
|
"grad_norm": 0.2387160062789917, |
|
"learning_rate": 7.103610675039247e-06, |
|
"loss": 1.1037, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.29120879120879123, |
|
"grad_norm": 0.2219192534685135, |
|
"learning_rate": 7.095761381475668e-06, |
|
"loss": 1.1385, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.29199372056514916, |
|
"grad_norm": 0.21481142938137054, |
|
"learning_rate": 7.087912087912089e-06, |
|
"loss": 1.1199, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.29277864992150704, |
|
"grad_norm": 0.2126331329345703, |
|
"learning_rate": 7.080062794348509e-06, |
|
"loss": 1.1346, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.29356357927786497, |
|
"grad_norm": 0.23164328932762146, |
|
"learning_rate": 7.07221350078493e-06, |
|
"loss": 1.1783, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.2943485086342229, |
|
"grad_norm": 0.23608547449111938, |
|
"learning_rate": 7.06436420722135e-06, |
|
"loss": 1.1549, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.29513343799058084, |
|
"grad_norm": 0.2102809101343155, |
|
"learning_rate": 7.056514913657771e-06, |
|
"loss": 1.144, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.29591836734693877, |
|
"grad_norm": 0.21914629638195038, |
|
"learning_rate": 7.048665620094192e-06, |
|
"loss": 1.1649, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.2967032967032967, |
|
"grad_norm": 0.22110004723072052, |
|
"learning_rate": 7.0408163265306125e-06, |
|
"loss": 1.1096, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.29748822605965464, |
|
"grad_norm": 0.23272880911827087, |
|
"learning_rate": 7.032967032967034e-06, |
|
"loss": 1.17, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.29827315541601257, |
|
"grad_norm": 0.2373398244380951, |
|
"learning_rate": 7.0251177394034545e-06, |
|
"loss": 1.1734, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.2990580847723705, |
|
"grad_norm": 0.20945794880390167, |
|
"learning_rate": 7.017268445839875e-06, |
|
"loss": 1.1192, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.29984301412872844, |
|
"grad_norm": 0.22829462587833405, |
|
"learning_rate": 7.0094191522762956e-06, |
|
"loss": 1.0966, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.30062794348508637, |
|
"grad_norm": 0.28657934069633484, |
|
"learning_rate": 7.001569858712717e-06, |
|
"loss": 1.0827, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.30141287284144425, |
|
"grad_norm": 0.2213151603937149, |
|
"learning_rate": 6.993720565149137e-06, |
|
"loss": 1.1452, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.3021978021978022, |
|
"grad_norm": 0.2172708660364151, |
|
"learning_rate": 6.985871271585558e-06, |
|
"loss": 1.1301, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.3029827315541601, |
|
"grad_norm": 0.21562695503234863, |
|
"learning_rate": 6.978021978021979e-06, |
|
"loss": 1.0969, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.30376766091051804, |
|
"grad_norm": 0.22482970356941223, |
|
"learning_rate": 6.970172684458399e-06, |
|
"loss": 1.0412, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.304552590266876, |
|
"grad_norm": 0.22429226338863373, |
|
"learning_rate": 6.96232339089482e-06, |
|
"loss": 1.1175, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.3053375196232339, |
|
"grad_norm": 0.306858628988266, |
|
"learning_rate": 6.954474097331241e-06, |
|
"loss": 1.1633, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.30612244897959184, |
|
"grad_norm": 0.22035177052021027, |
|
"learning_rate": 6.946624803767662e-06, |
|
"loss": 1.105, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.3069073783359498, |
|
"grad_norm": 0.20838768780231476, |
|
"learning_rate": 6.938775510204082e-06, |
|
"loss": 1.1126, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.3076923076923077, |
|
"grad_norm": 0.24387304484844208, |
|
"learning_rate": 6.930926216640504e-06, |
|
"loss": 1.0718, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.30847723704866564, |
|
"grad_norm": 0.20998525619506836, |
|
"learning_rate": 6.923076923076923e-06, |
|
"loss": 1.1408, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.3092621664050236, |
|
"grad_norm": 0.21903569996356964, |
|
"learning_rate": 6.915227629513345e-06, |
|
"loss": 1.1225, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.31004709576138145, |
|
"grad_norm": 0.21235527098178864, |
|
"learning_rate": 6.9073783359497645e-06, |
|
"loss": 1.0895, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.3108320251177394, |
|
"grad_norm": 0.24162974953651428, |
|
"learning_rate": 6.899529042386186e-06, |
|
"loss": 1.1487, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.3116169544740973, |
|
"grad_norm": 0.22563737630844116, |
|
"learning_rate": 6.8916797488226065e-06, |
|
"loss": 1.1442, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.31240188383045525, |
|
"grad_norm": 0.2697785794734955, |
|
"learning_rate": 6.883830455259027e-06, |
|
"loss": 1.2054, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.3131868131868132, |
|
"grad_norm": 0.20973092317581177, |
|
"learning_rate": 6.8759811616954476e-06, |
|
"loss": 1.1354, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.3139717425431711, |
|
"grad_norm": 0.21411223709583282, |
|
"learning_rate": 6.868131868131869e-06, |
|
"loss": 1.1232, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.31475667189952905, |
|
"grad_norm": 0.2229021191596985, |
|
"learning_rate": 6.8602825745682895e-06, |
|
"loss": 1.0909, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.315541601255887, |
|
"grad_norm": 0.25998759269714355, |
|
"learning_rate": 6.85243328100471e-06, |
|
"loss": 1.1639, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.3163265306122449, |
|
"grad_norm": 0.2209548056125641, |
|
"learning_rate": 6.8445839874411315e-06, |
|
"loss": 1.122, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.31711145996860285, |
|
"grad_norm": 0.2104836255311966, |
|
"learning_rate": 6.836734693877551e-06, |
|
"loss": 1.1052, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.3178963893249607, |
|
"grad_norm": 0.22340314090251923, |
|
"learning_rate": 6.828885400313973e-06, |
|
"loss": 1.1621, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.31868131868131866, |
|
"grad_norm": 0.20940294861793518, |
|
"learning_rate": 6.821036106750393e-06, |
|
"loss": 1.0895, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.3194662480376766, |
|
"grad_norm": 0.23529349267482758, |
|
"learning_rate": 6.813186813186814e-06, |
|
"loss": 1.1796, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.3202511773940345, |
|
"grad_norm": 0.22994717955589294, |
|
"learning_rate": 6.805337519623234e-06, |
|
"loss": 1.1736, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.32103610675039246, |
|
"grad_norm": 0.21799279749393463, |
|
"learning_rate": 6.797488226059656e-06, |
|
"loss": 1.0804, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.3218210361067504, |
|
"grad_norm": 0.22495627403259277, |
|
"learning_rate": 6.789638932496075e-06, |
|
"loss": 1.1605, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.3226059654631083, |
|
"grad_norm": 0.236924409866333, |
|
"learning_rate": 6.781789638932497e-06, |
|
"loss": 1.1081, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.32339089481946626, |
|
"grad_norm": 0.22784234583377838, |
|
"learning_rate": 6.773940345368918e-06, |
|
"loss": 1.092, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.3241758241758242, |
|
"grad_norm": 0.22424954175949097, |
|
"learning_rate": 6.766091051805338e-06, |
|
"loss": 1.1431, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.3249607535321821, |
|
"grad_norm": 0.22108778357505798, |
|
"learning_rate": 6.758241758241759e-06, |
|
"loss": 1.0865, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.32574568288854006, |
|
"grad_norm": 0.22631670534610748, |
|
"learning_rate": 6.750392464678179e-06, |
|
"loss": 1.1481, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.32653061224489793, |
|
"grad_norm": 0.23880915343761444, |
|
"learning_rate": 6.7425431711146e-06, |
|
"loss": 1.1503, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.32731554160125587, |
|
"grad_norm": 0.21916764974594116, |
|
"learning_rate": 6.734693877551021e-06, |
|
"loss": 1.1001, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.3281004709576138, |
|
"grad_norm": 0.22851671278476715, |
|
"learning_rate": 6.7268445839874415e-06, |
|
"loss": 1.0496, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.32888540031397173, |
|
"grad_norm": 0.2205754965543747, |
|
"learning_rate": 6.718995290423862e-06, |
|
"loss": 1.1324, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.32967032967032966, |
|
"grad_norm": 0.2369145303964615, |
|
"learning_rate": 6.7111459968602835e-06, |
|
"loss": 1.1575, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.3304552590266876, |
|
"grad_norm": 0.21567904949188232, |
|
"learning_rate": 6.703296703296703e-06, |
|
"loss": 1.1247, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.33124018838304553, |
|
"grad_norm": 0.22381411492824554, |
|
"learning_rate": 6.695447409733125e-06, |
|
"loss": 1.0777, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.33202511773940346, |
|
"grad_norm": 0.21957136690616608, |
|
"learning_rate": 6.687598116169546e-06, |
|
"loss": 1.0945, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.3328100470957614, |
|
"grad_norm": 0.22392071783542633, |
|
"learning_rate": 6.679748822605966e-06, |
|
"loss": 1.084, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.33359497645211933, |
|
"grad_norm": 0.21616291999816895, |
|
"learning_rate": 6.671899529042387e-06, |
|
"loss": 1.0848, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.33437990580847726, |
|
"grad_norm": 0.2230328768491745, |
|
"learning_rate": 6.664050235478807e-06, |
|
"loss": 1.1058, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.33516483516483514, |
|
"grad_norm": 0.21062208712100983, |
|
"learning_rate": 6.656200941915228e-06, |
|
"loss": 1.0867, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.3359497645211931, |
|
"grad_norm": 0.22089996933937073, |
|
"learning_rate": 6.648351648351649e-06, |
|
"loss": 1.1032, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.336734693877551, |
|
"grad_norm": 0.2328905314207077, |
|
"learning_rate": 6.64050235478807e-06, |
|
"loss": 1.1551, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.33751962323390894, |
|
"grad_norm": 0.23525848984718323, |
|
"learning_rate": 6.63265306122449e-06, |
|
"loss": 1.1529, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.33830455259026687, |
|
"grad_norm": 0.2343936562538147, |
|
"learning_rate": 6.624803767660911e-06, |
|
"loss": 1.1262, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.3390894819466248, |
|
"grad_norm": 0.2223115712404251, |
|
"learning_rate": 6.616954474097331e-06, |
|
"loss": 1.0855, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.33987441130298274, |
|
"grad_norm": 0.24484506249427795, |
|
"learning_rate": 6.609105180533752e-06, |
|
"loss": 1.0972, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.34065934065934067, |
|
"grad_norm": 0.22786875069141388, |
|
"learning_rate": 6.601255886970174e-06, |
|
"loss": 1.0784, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.3414442700156986, |
|
"grad_norm": 0.23298341035842896, |
|
"learning_rate": 6.5934065934065935e-06, |
|
"loss": 1.1613, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.34222919937205654, |
|
"grad_norm": 0.2376134991645813, |
|
"learning_rate": 6.585557299843015e-06, |
|
"loss": 1.0892, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.34301412872841447, |
|
"grad_norm": 0.23039846122264862, |
|
"learning_rate": 6.5777080062794355e-06, |
|
"loss": 1.1073, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.34379905808477235, |
|
"grad_norm": 0.23520535230636597, |
|
"learning_rate": 6.569858712715856e-06, |
|
"loss": 1.0612, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.3445839874411303, |
|
"grad_norm": 0.26625117659568787, |
|
"learning_rate": 6.562009419152277e-06, |
|
"loss": 1.0821, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.3453689167974882, |
|
"grad_norm": 0.22589145600795746, |
|
"learning_rate": 6.554160125588698e-06, |
|
"loss": 1.0766, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.34615384615384615, |
|
"grad_norm": 0.23303988575935364, |
|
"learning_rate": 6.546310832025118e-06, |
|
"loss": 1.1523, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.3469387755102041, |
|
"grad_norm": 0.24992690980434418, |
|
"learning_rate": 6.538461538461539e-06, |
|
"loss": 1.1378, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.347723704866562, |
|
"grad_norm": 0.23018966615200043, |
|
"learning_rate": 6.530612244897959e-06, |
|
"loss": 1.1121, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.34850863422291994, |
|
"grad_norm": 0.23048517107963562, |
|
"learning_rate": 6.52276295133438e-06, |
|
"loss": 1.1031, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.3492935635792779, |
|
"grad_norm": 0.239332914352417, |
|
"learning_rate": 6.514913657770802e-06, |
|
"loss": 1.1592, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.3500784929356358, |
|
"grad_norm": 0.236043319106102, |
|
"learning_rate": 6.507064364207221e-06, |
|
"loss": 1.082, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.35086342229199374, |
|
"grad_norm": 0.22672872245311737, |
|
"learning_rate": 6.499215070643643e-06, |
|
"loss": 1.1188, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.3516483516483517, |
|
"grad_norm": 0.23188328742980957, |
|
"learning_rate": 6.491365777080063e-06, |
|
"loss": 1.1368, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.35243328100470955, |
|
"grad_norm": 0.23719431459903717, |
|
"learning_rate": 6.483516483516485e-06, |
|
"loss": 1.1847, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.3532182103610675, |
|
"grad_norm": 0.23115630447864532, |
|
"learning_rate": 6.4756671899529044e-06, |
|
"loss": 1.1583, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.3540031397174254, |
|
"grad_norm": 0.22967442870140076, |
|
"learning_rate": 6.467817896389326e-06, |
|
"loss": 1.1319, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.35478806907378335, |
|
"grad_norm": 0.22553735971450806, |
|
"learning_rate": 6.4599686028257455e-06, |
|
"loss": 1.0739, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.3555729984301413, |
|
"grad_norm": 0.2251790314912796, |
|
"learning_rate": 6.452119309262167e-06, |
|
"loss": 1.1276, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.3563579277864992, |
|
"grad_norm": 0.22052745521068573, |
|
"learning_rate": 6.4442700156985875e-06, |
|
"loss": 1.1079, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.35714285714285715, |
|
"grad_norm": 0.2503249943256378, |
|
"learning_rate": 6.436420722135008e-06, |
|
"loss": 1.1554, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.3579277864992151, |
|
"grad_norm": 0.23954695463180542, |
|
"learning_rate": 6.4285714285714295e-06, |
|
"loss": 1.1269, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.358712715855573, |
|
"grad_norm": 0.2597452700138092, |
|
"learning_rate": 6.42072213500785e-06, |
|
"loss": 1.0967, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.35949764521193095, |
|
"grad_norm": 0.2288360893726349, |
|
"learning_rate": 6.4128728414442706e-06, |
|
"loss": 1.1297, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.3602825745682888, |
|
"grad_norm": 0.23878303170204163, |
|
"learning_rate": 6.405023547880691e-06, |
|
"loss": 1.1208, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.36106750392464676, |
|
"grad_norm": 0.22452662885189056, |
|
"learning_rate": 6.3971742543171125e-06, |
|
"loss": 1.0891, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.3618524332810047, |
|
"grad_norm": 0.23185127973556519, |
|
"learning_rate": 6.389324960753532e-06, |
|
"loss": 1.058, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.3626373626373626, |
|
"grad_norm": 0.23189175128936768, |
|
"learning_rate": 6.381475667189954e-06, |
|
"loss": 1.0812, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.36342229199372056, |
|
"grad_norm": 0.23450538516044617, |
|
"learning_rate": 6.373626373626373e-06, |
|
"loss": 1.1089, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.3642072213500785, |
|
"grad_norm": 0.22508233785629272, |
|
"learning_rate": 6.365777080062795e-06, |
|
"loss": 1.1062, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.3649921507064364, |
|
"grad_norm": 0.24207262694835663, |
|
"learning_rate": 6.357927786499215e-06, |
|
"loss": 1.1396, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.36577708006279436, |
|
"grad_norm": 0.2606029510498047, |
|
"learning_rate": 6.350078492935636e-06, |
|
"loss": 1.1261, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.3665620094191523, |
|
"grad_norm": 0.22962482273578644, |
|
"learning_rate": 6.342229199372057e-06, |
|
"loss": 1.1354, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.3673469387755102, |
|
"grad_norm": 0.3107023239135742, |
|
"learning_rate": 6.334379905808478e-06, |
|
"loss": 1.1735, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.36813186813186816, |
|
"grad_norm": 0.23643136024475098, |
|
"learning_rate": 6.326530612244899e-06, |
|
"loss": 1.086, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.36891679748822603, |
|
"grad_norm": 0.240147203207016, |
|
"learning_rate": 6.318681318681319e-06, |
|
"loss": 1.1356, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.36970172684458397, |
|
"grad_norm": 0.2438742220401764, |
|
"learning_rate": 6.31083202511774e-06, |
|
"loss": 1.1539, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.3704866562009419, |
|
"grad_norm": 0.2621842920780182, |
|
"learning_rate": 6.30298273155416e-06, |
|
"loss": 1.1623, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.37127158555729983, |
|
"grad_norm": 0.246039479970932, |
|
"learning_rate": 6.2951334379905815e-06, |
|
"loss": 1.1418, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.37205651491365777, |
|
"grad_norm": 0.2392723560333252, |
|
"learning_rate": 6.287284144427002e-06, |
|
"loss": 1.1692, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.3728414442700157, |
|
"grad_norm": 0.23897096514701843, |
|
"learning_rate": 6.279434850863423e-06, |
|
"loss": 1.111, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.37362637362637363, |
|
"grad_norm": 0.24796399474143982, |
|
"learning_rate": 6.271585557299843e-06, |
|
"loss": 1.0847, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.37441130298273156, |
|
"grad_norm": 0.33171799778938293, |
|
"learning_rate": 6.2637362637362645e-06, |
|
"loss": 1.226, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.3751962323390895, |
|
"grad_norm": 0.2655907869338989, |
|
"learning_rate": 6.255886970172685e-06, |
|
"loss": 1.0548, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.37598116169544743, |
|
"grad_norm": 0.2475586235523224, |
|
"learning_rate": 6.248037676609106e-06, |
|
"loss": 1.0914, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.37676609105180536, |
|
"grad_norm": 0.2385740876197815, |
|
"learning_rate": 6.240188383045527e-06, |
|
"loss": 1.0984, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.37755102040816324, |
|
"grad_norm": 0.27485817670822144, |
|
"learning_rate": 6.232339089481947e-06, |
|
"loss": 1.1461, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.3783359497645212, |
|
"grad_norm": 0.23348768055438995, |
|
"learning_rate": 6.224489795918368e-06, |
|
"loss": 1.0617, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.3791208791208791, |
|
"grad_norm": 0.23941084742546082, |
|
"learning_rate": 6.216640502354788e-06, |
|
"loss": 1.104, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.37990580847723704, |
|
"grad_norm": 0.22513240575790405, |
|
"learning_rate": 6.208791208791209e-06, |
|
"loss": 1.0354, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.38069073783359497, |
|
"grad_norm": 0.2495739459991455, |
|
"learning_rate": 6.20094191522763e-06, |
|
"loss": 1.1075, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.3814756671899529, |
|
"grad_norm": 0.23454253375530243, |
|
"learning_rate": 6.19309262166405e-06, |
|
"loss": 1.07, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.38226059654631084, |
|
"grad_norm": 0.2577785849571228, |
|
"learning_rate": 6.185243328100472e-06, |
|
"loss": 1.1229, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.38304552590266877, |
|
"grad_norm": 0.24680796265602112, |
|
"learning_rate": 6.177394034536892e-06, |
|
"loss": 1.1011, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.3838304552590267, |
|
"grad_norm": 0.23051689565181732, |
|
"learning_rate": 6.169544740973314e-06, |
|
"loss": 1.0953, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.38461538461538464, |
|
"grad_norm": 0.25582489371299744, |
|
"learning_rate": 6.1616954474097335e-06, |
|
"loss": 1.0825, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.38540031397174257, |
|
"grad_norm": 0.2379298061132431, |
|
"learning_rate": 6.153846153846155e-06, |
|
"loss": 1.0859, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.38618524332810045, |
|
"grad_norm": 0.2519659996032715, |
|
"learning_rate": 6.145996860282575e-06, |
|
"loss": 1.1339, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.3869701726844584, |
|
"grad_norm": 0.23378105461597443, |
|
"learning_rate": 6.138147566718996e-06, |
|
"loss": 1.0865, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.3877551020408163, |
|
"grad_norm": 0.24967290461063385, |
|
"learning_rate": 6.1302982731554165e-06, |
|
"loss": 1.1257, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.38854003139717425, |
|
"grad_norm": 0.2489061802625656, |
|
"learning_rate": 6.122448979591837e-06, |
|
"loss": 1.108, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.3893249607535322, |
|
"grad_norm": 0.24166519939899445, |
|
"learning_rate": 6.114599686028258e-06, |
|
"loss": 1.1026, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.3901098901098901, |
|
"grad_norm": 0.2487422674894333, |
|
"learning_rate": 6.106750392464679e-06, |
|
"loss": 1.144, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.39089481946624804, |
|
"grad_norm": 0.2340540736913681, |
|
"learning_rate": 6.0989010989011e-06, |
|
"loss": 1.0835, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.391679748822606, |
|
"grad_norm": 0.24633920192718506, |
|
"learning_rate": 6.09105180533752e-06, |
|
"loss": 1.0741, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.3924646781789639, |
|
"grad_norm": 0.23868292570114136, |
|
"learning_rate": 6.0832025117739416e-06, |
|
"loss": 1.0815, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.39324960753532184, |
|
"grad_norm": 0.2488425076007843, |
|
"learning_rate": 6.075353218210361e-06, |
|
"loss": 1.09, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.3940345368916798, |
|
"grad_norm": 0.22781234979629517, |
|
"learning_rate": 6.067503924646783e-06, |
|
"loss": 1.0947, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.39481946624803765, |
|
"grad_norm": 0.24967414140701294, |
|
"learning_rate": 6.059654631083202e-06, |
|
"loss": 1.1331, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.3956043956043956, |
|
"grad_norm": 0.2647199332714081, |
|
"learning_rate": 6.051805337519624e-06, |
|
"loss": 1.1436, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.3963893249607535, |
|
"grad_norm": 0.24577093124389648, |
|
"learning_rate": 6.043956043956044e-06, |
|
"loss": 1.0649, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.39717425431711145, |
|
"grad_norm": 0.23209503293037415, |
|
"learning_rate": 6.036106750392465e-06, |
|
"loss": 1.0968, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.3979591836734694, |
|
"grad_norm": 0.23621460795402527, |
|
"learning_rate": 6.0282574568288855e-06, |
|
"loss": 1.0858, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.3987441130298273, |
|
"grad_norm": 0.23415528237819672, |
|
"learning_rate": 6.020408163265307e-06, |
|
"loss": 1.039, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.39952904238618525, |
|
"grad_norm": 0.24313485622406006, |
|
"learning_rate": 6.012558869701728e-06, |
|
"loss": 1.1233, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.4003139717425432, |
|
"grad_norm": 0.24086087942123413, |
|
"learning_rate": 6.004709576138148e-06, |
|
"loss": 1.0779, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.4010989010989011, |
|
"grad_norm": 0.22834287583827972, |
|
"learning_rate": 5.996860282574569e-06, |
|
"loss": 1.0448, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.40188383045525905, |
|
"grad_norm": 0.2525885999202728, |
|
"learning_rate": 5.989010989010989e-06, |
|
"loss": 1.0905, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.4026687598116169, |
|
"grad_norm": 0.26272568106651306, |
|
"learning_rate": 5.9811616954474105e-06, |
|
"loss": 1.114, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.40345368916797486, |
|
"grad_norm": 0.2448360174894333, |
|
"learning_rate": 5.973312401883831e-06, |
|
"loss": 1.0682, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.4042386185243328, |
|
"grad_norm": 0.24594442546367645, |
|
"learning_rate": 5.965463108320252e-06, |
|
"loss": 1.1067, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.4050235478806907, |
|
"grad_norm": 0.23227113485336304, |
|
"learning_rate": 5.957613814756672e-06, |
|
"loss": 1.0592, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.40580847723704866, |
|
"grad_norm": 0.24381886422634125, |
|
"learning_rate": 5.949764521193094e-06, |
|
"loss": 1.0932, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.4065934065934066, |
|
"grad_norm": 0.2468293309211731, |
|
"learning_rate": 5.941915227629513e-06, |
|
"loss": 1.0975, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.4073783359497645, |
|
"grad_norm": 0.24225494265556335, |
|
"learning_rate": 5.934065934065935e-06, |
|
"loss": 1.095, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.40816326530612246, |
|
"grad_norm": 0.2511250972747803, |
|
"learning_rate": 5.926216640502356e-06, |
|
"loss": 1.1268, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.4089481946624804, |
|
"grad_norm": 0.25592711567878723, |
|
"learning_rate": 5.918367346938776e-06, |
|
"loss": 1.0655, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.4097331240188383, |
|
"grad_norm": 0.2539741098880768, |
|
"learning_rate": 5.910518053375197e-06, |
|
"loss": 1.0909, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.41051805337519626, |
|
"grad_norm": 0.2586928904056549, |
|
"learning_rate": 5.902668759811617e-06, |
|
"loss": 1.102, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.41130298273155413, |
|
"grad_norm": 0.2530655860900879, |
|
"learning_rate": 5.894819466248038e-06, |
|
"loss": 1.1278, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.41208791208791207, |
|
"grad_norm": 0.2644376754760742, |
|
"learning_rate": 5.886970172684459e-06, |
|
"loss": 1.0853, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.41287284144427, |
|
"grad_norm": 0.2579070031642914, |
|
"learning_rate": 5.8791208791208794e-06, |
|
"loss": 1.0258, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.41365777080062793, |
|
"grad_norm": 0.2522946894168854, |
|
"learning_rate": 5.8712715855573e-06, |
|
"loss": 1.0873, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.41444270015698587, |
|
"grad_norm": 0.24585580825805664, |
|
"learning_rate": 5.863422291993721e-06, |
|
"loss": 1.0685, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.4152276295133438, |
|
"grad_norm": 0.23922014236450195, |
|
"learning_rate": 5.855572998430141e-06, |
|
"loss": 1.0501, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.41601255886970173, |
|
"grad_norm": 0.3307192325592041, |
|
"learning_rate": 5.8477237048665625e-06, |
|
"loss": 1.098, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.41679748822605966, |
|
"grad_norm": 0.26279258728027344, |
|
"learning_rate": 5.839874411302984e-06, |
|
"loss": 1.1451, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.4175824175824176, |
|
"grad_norm": 0.24915215373039246, |
|
"learning_rate": 5.832025117739404e-06, |
|
"loss": 1.0702, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.41836734693877553, |
|
"grad_norm": 0.3191397488117218, |
|
"learning_rate": 5.824175824175825e-06, |
|
"loss": 1.1355, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.41915227629513346, |
|
"grad_norm": 0.24395009875297546, |
|
"learning_rate": 5.816326530612246e-06, |
|
"loss": 1.0693, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.41993720565149134, |
|
"grad_norm": 0.2642136812210083, |
|
"learning_rate": 5.808477237048666e-06, |
|
"loss": 1.0774, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.4207221350078493, |
|
"grad_norm": 0.24453726410865784, |
|
"learning_rate": 5.800627943485087e-06, |
|
"loss": 1.0767, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.4215070643642072, |
|
"grad_norm": 0.381024569272995, |
|
"learning_rate": 5.792778649921508e-06, |
|
"loss": 1.1438, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.42229199372056514, |
|
"grad_norm": 0.30179327726364136, |
|
"learning_rate": 5.784929356357928e-06, |
|
"loss": 1.1929, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.4230769230769231, |
|
"grad_norm": 0.24929693341255188, |
|
"learning_rate": 5.777080062794349e-06, |
|
"loss": 1.1482, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.423861852433281, |
|
"grad_norm": 0.26368796825408936, |
|
"learning_rate": 5.769230769230769e-06, |
|
"loss": 1.0321, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.42464678178963894, |
|
"grad_norm": 0.2380351573228836, |
|
"learning_rate": 5.76138147566719e-06, |
|
"loss": 1.1042, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.42543171114599687, |
|
"grad_norm": 0.25883641839027405, |
|
"learning_rate": 5.753532182103612e-06, |
|
"loss": 1.0072, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.4262166405023548, |
|
"grad_norm": 0.23825643956661224, |
|
"learning_rate": 5.7456828885400314e-06, |
|
"loss": 1.0884, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.42700156985871274, |
|
"grad_norm": 0.3005066215991974, |
|
"learning_rate": 5.737833594976453e-06, |
|
"loss": 1.0975, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.42778649921507067, |
|
"grad_norm": 0.24875208735466003, |
|
"learning_rate": 5.729984301412873e-06, |
|
"loss": 1.1003, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.42857142857142855, |
|
"grad_norm": 0.350685715675354, |
|
"learning_rate": 5.722135007849294e-06, |
|
"loss": 1.0723, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.4293563579277865, |
|
"grad_norm": 0.2770320177078247, |
|
"learning_rate": 5.7142857142857145e-06, |
|
"loss": 1.1224, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.4301412872841444, |
|
"grad_norm": 0.24251538515090942, |
|
"learning_rate": 5.706436420722136e-06, |
|
"loss": 1.0548, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.43092621664050235, |
|
"grad_norm": 0.24872933328151703, |
|
"learning_rate": 5.698587127158556e-06, |
|
"loss": 1.0798, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.4317111459968603, |
|
"grad_norm": 0.2544524669647217, |
|
"learning_rate": 5.690737833594977e-06, |
|
"loss": 1.1251, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.4324960753532182, |
|
"grad_norm": 0.24090701341629028, |
|
"learning_rate": 5.682888540031397e-06, |
|
"loss": 1.1367, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.43328100470957615, |
|
"grad_norm": 0.23915883898735046, |
|
"learning_rate": 5.675039246467818e-06, |
|
"loss": 1.0647, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.4340659340659341, |
|
"grad_norm": 0.2448183000087738, |
|
"learning_rate": 5.6671899529042395e-06, |
|
"loss": 1.0755, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.434850863422292, |
|
"grad_norm": 0.25648248195648193, |
|
"learning_rate": 5.65934065934066e-06, |
|
"loss": 1.0525, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.43563579277864994, |
|
"grad_norm": 0.23780952394008636, |
|
"learning_rate": 5.651491365777081e-06, |
|
"loss": 1.0902, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.4364207221350079, |
|
"grad_norm": 0.2836390733718872, |
|
"learning_rate": 5.643642072213501e-06, |
|
"loss": 1.0601, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.43720565149136575, |
|
"grad_norm": 0.24821418523788452, |
|
"learning_rate": 5.635792778649923e-06, |
|
"loss": 1.083, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.4379905808477237, |
|
"grad_norm": 0.25154179334640503, |
|
"learning_rate": 5.627943485086342e-06, |
|
"loss": 1.1134, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.4387755102040816, |
|
"grad_norm": 0.26084184646606445, |
|
"learning_rate": 5.620094191522764e-06, |
|
"loss": 1.111, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.43956043956043955, |
|
"grad_norm": 0.2489292174577713, |
|
"learning_rate": 5.6122448979591834e-06, |
|
"loss": 1.1249, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.4403453689167975, |
|
"grad_norm": 0.2615135610103607, |
|
"learning_rate": 5.604395604395605e-06, |
|
"loss": 1.1283, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.4411302982731554, |
|
"grad_norm": 0.2750528156757355, |
|
"learning_rate": 5.596546310832025e-06, |
|
"loss": 1.0626, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.44191522762951335, |
|
"grad_norm": 0.2562703788280487, |
|
"learning_rate": 5.588697017268446e-06, |
|
"loss": 1.0732, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.4427001569858713, |
|
"grad_norm": 0.25479015707969666, |
|
"learning_rate": 5.580847723704867e-06, |
|
"loss": 1.0509, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.4434850863422292, |
|
"grad_norm": 0.30177152156829834, |
|
"learning_rate": 5.572998430141288e-06, |
|
"loss": 1.0856, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.44427001569858715, |
|
"grad_norm": 0.2459626942873001, |
|
"learning_rate": 5.5651491365777085e-06, |
|
"loss": 1.0828, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.44505494505494503, |
|
"grad_norm": 0.2764773964881897, |
|
"learning_rate": 5.557299843014129e-06, |
|
"loss": 1.1437, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.44583987441130296, |
|
"grad_norm": 0.24312351644039154, |
|
"learning_rate": 5.5494505494505504e-06, |
|
"loss": 1.0679, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.4466248037676609, |
|
"grad_norm": 0.5099506974220276, |
|
"learning_rate": 5.54160125588697e-06, |
|
"loss": 1.1154, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.4474097331240188, |
|
"grad_norm": 0.24318666756153107, |
|
"learning_rate": 5.5337519623233915e-06, |
|
"loss": 1.0725, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.44819466248037676, |
|
"grad_norm": 0.2769014835357666, |
|
"learning_rate": 5.525902668759811e-06, |
|
"loss": 1.1424, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.4489795918367347, |
|
"grad_norm": 0.2566874921321869, |
|
"learning_rate": 5.518053375196233e-06, |
|
"loss": 1.1231, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.4497645211930926, |
|
"grad_norm": 0.27724429965019226, |
|
"learning_rate": 5.510204081632653e-06, |
|
"loss": 1.105, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.45054945054945056, |
|
"grad_norm": 0.28734901547431946, |
|
"learning_rate": 5.502354788069074e-06, |
|
"loss": 1.1076, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.4513343799058085, |
|
"grad_norm": 0.25727584958076477, |
|
"learning_rate": 5.494505494505495e-06, |
|
"loss": 1.0624, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.4521193092621664, |
|
"grad_norm": 0.2726932764053345, |
|
"learning_rate": 5.486656200941916e-06, |
|
"loss": 1.0954, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.45290423861852436, |
|
"grad_norm": 0.249312624335289, |
|
"learning_rate": 5.478806907378337e-06, |
|
"loss": 1.1038, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.45368916797488223, |
|
"grad_norm": 0.2610965669155121, |
|
"learning_rate": 5.470957613814757e-06, |
|
"loss": 1.1033, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.45447409733124017, |
|
"grad_norm": 0.2687593400478363, |
|
"learning_rate": 5.463108320251178e-06, |
|
"loss": 1.1013, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.4552590266875981, |
|
"grad_norm": 0.2867527902126312, |
|
"learning_rate": 5.455259026687598e-06, |
|
"loss": 1.104, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.45604395604395603, |
|
"grad_norm": 0.2627805769443512, |
|
"learning_rate": 5.447409733124019e-06, |
|
"loss": 1.0958, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.45682888540031397, |
|
"grad_norm": 0.25707143545150757, |
|
"learning_rate": 5.43956043956044e-06, |
|
"loss": 1.1073, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.4576138147566719, |
|
"grad_norm": 0.26457729935646057, |
|
"learning_rate": 5.4317111459968605e-06, |
|
"loss": 1.1379, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.45839874411302983, |
|
"grad_norm": 0.25039640069007874, |
|
"learning_rate": 5.423861852433281e-06, |
|
"loss": 1.1107, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.45918367346938777, |
|
"grad_norm": 0.2630631923675537, |
|
"learning_rate": 5.4160125588697024e-06, |
|
"loss": 1.0336, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.4599686028257457, |
|
"grad_norm": 0.2637465298175812, |
|
"learning_rate": 5.408163265306123e-06, |
|
"loss": 1.1183, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.46075353218210363, |
|
"grad_norm": 0.2565993368625641, |
|
"learning_rate": 5.4003139717425436e-06, |
|
"loss": 1.0769, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.46153846153846156, |
|
"grad_norm": 0.2511787414550781, |
|
"learning_rate": 5.392464678178965e-06, |
|
"loss": 1.1005, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.46232339089481944, |
|
"grad_norm": 0.24590060114860535, |
|
"learning_rate": 5.384615384615385e-06, |
|
"loss": 1.1103, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.4631083202511774, |
|
"grad_norm": 0.26895543932914734, |
|
"learning_rate": 5.376766091051806e-06, |
|
"loss": 1.0971, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.4638932496075353, |
|
"grad_norm": 0.25688987970352173, |
|
"learning_rate": 5.368916797488226e-06, |
|
"loss": 1.0973, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.46467817896389324, |
|
"grad_norm": 0.2657226026058197, |
|
"learning_rate": 5.361067503924647e-06, |
|
"loss": 1.124, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.4654631083202512, |
|
"grad_norm": 0.2537216544151306, |
|
"learning_rate": 5.353218210361068e-06, |
|
"loss": 1.0958, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.4662480376766091, |
|
"grad_norm": 0.32557958364486694, |
|
"learning_rate": 5.345368916797488e-06, |
|
"loss": 1.1469, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.46703296703296704, |
|
"grad_norm": 0.25399667024612427, |
|
"learning_rate": 5.33751962323391e-06, |
|
"loss": 1.0587, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.46781789638932497, |
|
"grad_norm": 0.26794350147247314, |
|
"learning_rate": 5.32967032967033e-06, |
|
"loss": 1.058, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.4686028257456829, |
|
"grad_norm": 0.2695688009262085, |
|
"learning_rate": 5.321821036106752e-06, |
|
"loss": 1.1169, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.46938775510204084, |
|
"grad_norm": 0.2774251103401184, |
|
"learning_rate": 5.313971742543171e-06, |
|
"loss": 1.1368, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.47017268445839877, |
|
"grad_norm": 0.25040146708488464, |
|
"learning_rate": 5.306122448979593e-06, |
|
"loss": 1.0831, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.47095761381475665, |
|
"grad_norm": 0.2619199752807617, |
|
"learning_rate": 5.2982731554160125e-06, |
|
"loss": 1.0354, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.4717425431711146, |
|
"grad_norm": 0.25245144963264465, |
|
"learning_rate": 5.290423861852434e-06, |
|
"loss": 1.0487, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.4725274725274725, |
|
"grad_norm": 0.25916045904159546, |
|
"learning_rate": 5.2825745682888544e-06, |
|
"loss": 1.0987, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.47331240188383045, |
|
"grad_norm": 0.2599029541015625, |
|
"learning_rate": 5.274725274725275e-06, |
|
"loss": 1.0945, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.4740973312401884, |
|
"grad_norm": 0.27398571372032166, |
|
"learning_rate": 5.2668759811616956e-06, |
|
"loss": 1.0602, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.4748822605965463, |
|
"grad_norm": 0.24853894114494324, |
|
"learning_rate": 5.259026687598117e-06, |
|
"loss": 1.0169, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.47566718995290425, |
|
"grad_norm": 0.24999196827411652, |
|
"learning_rate": 5.2511773940345375e-06, |
|
"loss": 1.0557, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.4764521193092622, |
|
"grad_norm": 0.2587362229824066, |
|
"learning_rate": 5.243328100470958e-06, |
|
"loss": 1.0832, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.4772370486656201, |
|
"grad_norm": 0.255636990070343, |
|
"learning_rate": 5.2354788069073795e-06, |
|
"loss": 1.0378, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.47802197802197804, |
|
"grad_norm": 0.2529115080833435, |
|
"learning_rate": 5.227629513343799e-06, |
|
"loss": 1.0544, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.478806907378336, |
|
"grad_norm": 0.3072440028190613, |
|
"learning_rate": 5.219780219780221e-06, |
|
"loss": 1.1386, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.47959183673469385, |
|
"grad_norm": 0.24107390642166138, |
|
"learning_rate": 5.21193092621664e-06, |
|
"loss": 1.0658, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.4803767660910518, |
|
"grad_norm": 0.2693899869918823, |
|
"learning_rate": 5.204081632653062e-06, |
|
"loss": 1.0966, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.4811616954474097, |
|
"grad_norm": 0.2495475560426712, |
|
"learning_rate": 5.196232339089482e-06, |
|
"loss": 1.087, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.48194662480376765, |
|
"grad_norm": 0.2680012285709381, |
|
"learning_rate": 5.188383045525903e-06, |
|
"loss": 1.092, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.4827315541601256, |
|
"grad_norm": 0.2574861943721771, |
|
"learning_rate": 5.180533751962323e-06, |
|
"loss": 1.089, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.4835164835164835, |
|
"grad_norm": 0.2980441153049469, |
|
"learning_rate": 5.172684458398745e-06, |
|
"loss": 1.0895, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.48430141287284145, |
|
"grad_norm": 0.2533935606479645, |
|
"learning_rate": 5.164835164835166e-06, |
|
"loss": 1.0627, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.4850863422291994, |
|
"grad_norm": 0.2828797399997711, |
|
"learning_rate": 5.156985871271586e-06, |
|
"loss": 1.0874, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.4858712715855573, |
|
"grad_norm": 0.26118507981300354, |
|
"learning_rate": 5.149136577708007e-06, |
|
"loss": 1.0304, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.48665620094191525, |
|
"grad_norm": 0.277045339345932, |
|
"learning_rate": 5.141287284144427e-06, |
|
"loss": 1.0492, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.48744113029827313, |
|
"grad_norm": 0.2713971734046936, |
|
"learning_rate": 5.133437990580848e-06, |
|
"loss": 1.1037, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.48822605965463106, |
|
"grad_norm": 0.27905189990997314, |
|
"learning_rate": 5.125588697017269e-06, |
|
"loss": 1.0899, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.489010989010989, |
|
"grad_norm": 0.2574610710144043, |
|
"learning_rate": 5.1177394034536895e-06, |
|
"loss": 1.0842, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.4897959183673469, |
|
"grad_norm": 0.25002968311309814, |
|
"learning_rate": 5.10989010989011e-06, |
|
"loss": 1.0707, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.49058084772370486, |
|
"grad_norm": 0.25703486800193787, |
|
"learning_rate": 5.1020408163265315e-06, |
|
"loss": 1.0427, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.4913657770800628, |
|
"grad_norm": 0.26097506284713745, |
|
"learning_rate": 5.094191522762951e-06, |
|
"loss": 1.0554, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.4921507064364207, |
|
"grad_norm": 0.25719407200813293, |
|
"learning_rate": 5.086342229199373e-06, |
|
"loss": 1.0612, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.49293563579277866, |
|
"grad_norm": 0.25571000576019287, |
|
"learning_rate": 5.078492935635794e-06, |
|
"loss": 1.0471, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.4937205651491366, |
|
"grad_norm": 0.2634165287017822, |
|
"learning_rate": 5.070643642072214e-06, |
|
"loss": 1.0811, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.4945054945054945, |
|
"grad_norm": 0.247939795255661, |
|
"learning_rate": 5.062794348508635e-06, |
|
"loss": 1.0038, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.49529042386185246, |
|
"grad_norm": 0.27619796991348267, |
|
"learning_rate": 5.054945054945055e-06, |
|
"loss": 1.1057, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.49607535321821034, |
|
"grad_norm": 0.26260074973106384, |
|
"learning_rate": 5.047095761381476e-06, |
|
"loss": 1.1193, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.49686028257456827, |
|
"grad_norm": 0.258558988571167, |
|
"learning_rate": 5.039246467817897e-06, |
|
"loss": 1.0687, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.4976452119309262, |
|
"grad_norm": 0.26512664556503296, |
|
"learning_rate": 5.031397174254317e-06, |
|
"loss": 1.0561, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.49843014128728413, |
|
"grad_norm": 0.2602575421333313, |
|
"learning_rate": 5.023547880690738e-06, |
|
"loss": 1.0438, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.49921507064364207, |
|
"grad_norm": 0.2699330151081085, |
|
"learning_rate": 5.015698587127159e-06, |
|
"loss": 1.0989, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.27960050106048584, |
|
"learning_rate": 5.007849293563579e-06, |
|
"loss": 1.0354, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.5007849293563579, |
|
"grad_norm": 0.287338525056839, |
|
"learning_rate": 5e-06, |
|
"loss": 1.1538, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.5015698587127159, |
|
"grad_norm": 0.26822763681411743, |
|
"learning_rate": 4.992150706436421e-06, |
|
"loss": 1.0374, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.5023547880690737, |
|
"grad_norm": 0.2564684748649597, |
|
"learning_rate": 4.9843014128728415e-06, |
|
"loss": 1.0611, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.5031397174254317, |
|
"grad_norm": 0.2621293365955353, |
|
"learning_rate": 4.976452119309262e-06, |
|
"loss": 1.0946, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.5039246467817896, |
|
"grad_norm": 0.2555181086063385, |
|
"learning_rate": 4.9686028257456835e-06, |
|
"loss": 1.0753, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.5047095761381476, |
|
"grad_norm": 0.27201956510543823, |
|
"learning_rate": 4.960753532182104e-06, |
|
"loss": 1.0945, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.5054945054945055, |
|
"grad_norm": 0.2534913420677185, |
|
"learning_rate": 4.952904238618525e-06, |
|
"loss": 1.0746, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.5062794348508635, |
|
"grad_norm": 0.25126805901527405, |
|
"learning_rate": 4.945054945054946e-06, |
|
"loss": 1.0589, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.5070643642072213, |
|
"grad_norm": 0.27167466282844543, |
|
"learning_rate": 4.9372056514913666e-06, |
|
"loss": 1.1345, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.5078492935635793, |
|
"grad_norm": 0.25426211953163147, |
|
"learning_rate": 4.929356357927787e-06, |
|
"loss": 1.0779, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.5086342229199372, |
|
"grad_norm": 0.2680758237838745, |
|
"learning_rate": 4.921507064364208e-06, |
|
"loss": 1.0908, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.5094191522762951, |
|
"grad_norm": 0.26956409215927124, |
|
"learning_rate": 4.913657770800628e-06, |
|
"loss": 1.0492, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.5102040816326531, |
|
"grad_norm": 0.2579835057258606, |
|
"learning_rate": 4.905808477237049e-06, |
|
"loss": 1.0819, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.510989010989011, |
|
"grad_norm": 0.2698979675769806, |
|
"learning_rate": 4.897959183673469e-06, |
|
"loss": 1.069, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.5117739403453689, |
|
"grad_norm": 0.27058646082878113, |
|
"learning_rate": 4.890109890109891e-06, |
|
"loss": 1.0652, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.5125588697017268, |
|
"grad_norm": 0.28998565673828125, |
|
"learning_rate": 4.882260596546311e-06, |
|
"loss": 1.1569, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.5133437990580848, |
|
"grad_norm": 0.2735849916934967, |
|
"learning_rate": 4.874411302982732e-06, |
|
"loss": 1.084, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.5141287284144427, |
|
"grad_norm": 0.2719517648220062, |
|
"learning_rate": 4.866562009419153e-06, |
|
"loss": 1.0997, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.5149136577708007, |
|
"grad_norm": 0.2657535672187805, |
|
"learning_rate": 4.858712715855574e-06, |
|
"loss": 1.0977, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.5156985871271585, |
|
"grad_norm": 0.2546514570713043, |
|
"learning_rate": 4.850863422291994e-06, |
|
"loss": 1.0897, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.5164835164835165, |
|
"grad_norm": 0.2730329632759094, |
|
"learning_rate": 4.843014128728415e-06, |
|
"loss": 1.1082, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.5172684458398744, |
|
"grad_norm": 0.2657451331615448, |
|
"learning_rate": 4.8351648351648355e-06, |
|
"loss": 1.0823, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.5180533751962323, |
|
"grad_norm": 0.28653398156166077, |
|
"learning_rate": 4.827315541601256e-06, |
|
"loss": 1.0555, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.5188383045525903, |
|
"grad_norm": 0.2710787355899811, |
|
"learning_rate": 4.819466248037677e-06, |
|
"loss": 1.0385, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.5196232339089482, |
|
"grad_norm": 0.26727956533432007, |
|
"learning_rate": 4.811616954474098e-06, |
|
"loss": 1.068, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.5204081632653061, |
|
"grad_norm": 0.25900620222091675, |
|
"learning_rate": 4.8037676609105186e-06, |
|
"loss": 1.0889, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.521193092621664, |
|
"grad_norm": 0.26514261960983276, |
|
"learning_rate": 4.795918367346939e-06, |
|
"loss": 1.0978, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.521978021978022, |
|
"grad_norm": 0.29534074664115906, |
|
"learning_rate": 4.7880690737833605e-06, |
|
"loss": 1.0289, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.5227629513343799, |
|
"grad_norm": 0.2783236801624298, |
|
"learning_rate": 4.780219780219781e-06, |
|
"loss": 1.1002, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.5235478806907379, |
|
"grad_norm": 0.2634176015853882, |
|
"learning_rate": 4.772370486656202e-06, |
|
"loss": 1.0518, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.5243328100470958, |
|
"grad_norm": 0.27170518040657043, |
|
"learning_rate": 4.764521193092622e-06, |
|
"loss": 1.0823, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.5251177394034537, |
|
"grad_norm": 0.26880884170532227, |
|
"learning_rate": 4.756671899529043e-06, |
|
"loss": 1.0384, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.5259026687598116, |
|
"grad_norm": 0.27249330282211304, |
|
"learning_rate": 4.748822605965463e-06, |
|
"loss": 1.0706, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.5266875981161695, |
|
"grad_norm": 0.2680225074291229, |
|
"learning_rate": 4.740973312401884e-06, |
|
"loss": 1.0794, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.5274725274725275, |
|
"grad_norm": 0.27644068002700806, |
|
"learning_rate": 4.733124018838305e-06, |
|
"loss": 1.1037, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.5282574568288854, |
|
"grad_norm": 0.2654259204864502, |
|
"learning_rate": 4.725274725274726e-06, |
|
"loss": 1.0524, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.5290423861852434, |
|
"grad_norm": 0.2548593282699585, |
|
"learning_rate": 4.717425431711146e-06, |
|
"loss": 1.0596, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.5298273155416012, |
|
"grad_norm": 0.262279748916626, |
|
"learning_rate": 4.709576138147567e-06, |
|
"loss": 1.0731, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.5306122448979592, |
|
"grad_norm": 0.3365892767906189, |
|
"learning_rate": 4.701726844583988e-06, |
|
"loss": 1.1082, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.5313971742543171, |
|
"grad_norm": 0.25070279836654663, |
|
"learning_rate": 4.693877551020409e-06, |
|
"loss": 1.0554, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.5321821036106751, |
|
"grad_norm": 0.3416726291179657, |
|
"learning_rate": 4.6860282574568294e-06, |
|
"loss": 1.0501, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.532967032967033, |
|
"grad_norm": 0.2714819610118866, |
|
"learning_rate": 4.67817896389325e-06, |
|
"loss": 1.0317, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.533751962323391, |
|
"grad_norm": 0.2663348615169525, |
|
"learning_rate": 4.6703296703296706e-06, |
|
"loss": 1.0699, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.5345368916797488, |
|
"grad_norm": 0.26617753505706787, |
|
"learning_rate": 4.662480376766091e-06, |
|
"loss": 1.0755, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.5353218210361067, |
|
"grad_norm": 0.25647595524787903, |
|
"learning_rate": 4.6546310832025125e-06, |
|
"loss": 1.0746, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.5361067503924647, |
|
"grad_norm": 0.2799610197544098, |
|
"learning_rate": 4.646781789638933e-06, |
|
"loss": 1.0977, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.5368916797488226, |
|
"grad_norm": 0.26751118898391724, |
|
"learning_rate": 4.638932496075354e-06, |
|
"loss": 1.0633, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.5376766091051806, |
|
"grad_norm": 0.28023260831832886, |
|
"learning_rate": 4.631083202511774e-06, |
|
"loss": 1.0927, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.5384615384615384, |
|
"grad_norm": 0.27010107040405273, |
|
"learning_rate": 4.623233908948195e-06, |
|
"loss": 1.0869, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.5392464678178964, |
|
"grad_norm": 0.2625042796134949, |
|
"learning_rate": 4.615384615384616e-06, |
|
"loss": 1.0615, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.5400313971742543, |
|
"grad_norm": 0.2663073241710663, |
|
"learning_rate": 4.607535321821037e-06, |
|
"loss": 1.0473, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.5408163265306123, |
|
"grad_norm": 0.26662012934684753, |
|
"learning_rate": 4.599686028257457e-06, |
|
"loss": 1.0611, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.5416012558869702, |
|
"grad_norm": 0.2595308721065521, |
|
"learning_rate": 4.591836734693878e-06, |
|
"loss": 1.0645, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.542386185243328, |
|
"grad_norm": 0.286700040102005, |
|
"learning_rate": 4.583987441130298e-06, |
|
"loss": 1.0211, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.543171114599686, |
|
"grad_norm": 0.2713211476802826, |
|
"learning_rate": 4.57613814756672e-06, |
|
"loss": 1.0621, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.5439560439560439, |
|
"grad_norm": 0.2682141363620758, |
|
"learning_rate": 4.56828885400314e-06, |
|
"loss": 1.0571, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.5447409733124019, |
|
"grad_norm": 0.2777917683124542, |
|
"learning_rate": 4.560439560439561e-06, |
|
"loss": 1.0733, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.5455259026687598, |
|
"grad_norm": 0.25854945182800293, |
|
"learning_rate": 4.5525902668759815e-06, |
|
"loss": 1.0926, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.5463108320251178, |
|
"grad_norm": 0.2838626801967621, |
|
"learning_rate": 4.544740973312402e-06, |
|
"loss": 1.0707, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.5470957613814756, |
|
"grad_norm": 0.26979881525039673, |
|
"learning_rate": 4.5368916797488226e-06, |
|
"loss": 1.0949, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.5478806907378336, |
|
"grad_norm": 0.2945154309272766, |
|
"learning_rate": 4.529042386185244e-06, |
|
"loss": 1.1497, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.5486656200941915, |
|
"grad_norm": 0.25463616847991943, |
|
"learning_rate": 4.5211930926216645e-06, |
|
"loss": 1.0762, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.5494505494505495, |
|
"grad_norm": 0.2613489031791687, |
|
"learning_rate": 4.513343799058085e-06, |
|
"loss": 1.0629, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.5502354788069074, |
|
"grad_norm": 0.2718147039413452, |
|
"learning_rate": 4.505494505494506e-06, |
|
"loss": 1.1294, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.5510204081632653, |
|
"grad_norm": 0.2775886654853821, |
|
"learning_rate": 4.497645211930927e-06, |
|
"loss": 1.0303, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.5518053375196232, |
|
"grad_norm": 0.2916508913040161, |
|
"learning_rate": 4.489795918367348e-06, |
|
"loss": 1.0797, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.5525902668759811, |
|
"grad_norm": 0.2996635138988495, |
|
"learning_rate": 4.481946624803768e-06, |
|
"loss": 1.0696, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.5533751962323391, |
|
"grad_norm": 0.31456199288368225, |
|
"learning_rate": 4.474097331240189e-06, |
|
"loss": 1.0751, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.554160125588697, |
|
"grad_norm": 0.26043495535850525, |
|
"learning_rate": 4.466248037676609e-06, |
|
"loss": 1.0588, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.554945054945055, |
|
"grad_norm": 0.27495190501213074, |
|
"learning_rate": 4.45839874411303e-06, |
|
"loss": 1.0479, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.5557299843014128, |
|
"grad_norm": 0.2717645764350891, |
|
"learning_rate": 4.45054945054945e-06, |
|
"loss": 1.0718, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.5565149136577708, |
|
"grad_norm": 0.29588571190834045, |
|
"learning_rate": 4.442700156985872e-06, |
|
"loss": 1.0585, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.5572998430141287, |
|
"grad_norm": 0.2885316014289856, |
|
"learning_rate": 4.434850863422292e-06, |
|
"loss": 1.0565, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.5580847723704867, |
|
"grad_norm": 0.2899274230003357, |
|
"learning_rate": 4.427001569858713e-06, |
|
"loss": 1.0994, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.5588697017268446, |
|
"grad_norm": 0.27042970061302185, |
|
"learning_rate": 4.419152276295134e-06, |
|
"loss": 1.0642, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.5596546310832025, |
|
"grad_norm": 0.2861323654651642, |
|
"learning_rate": 4.411302982731555e-06, |
|
"loss": 1.0588, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.5604395604395604, |
|
"grad_norm": 0.2717403471469879, |
|
"learning_rate": 4.403453689167975e-06, |
|
"loss": 1.0684, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.5612244897959183, |
|
"grad_norm": 0.2660480737686157, |
|
"learning_rate": 4.395604395604396e-06, |
|
"loss": 1.0709, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.5620094191522763, |
|
"grad_norm": 0.27813735604286194, |
|
"learning_rate": 4.3877551020408165e-06, |
|
"loss": 1.0424, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.5627943485086342, |
|
"grad_norm": 0.27436062693595886, |
|
"learning_rate": 4.379905808477237e-06, |
|
"loss": 1.118, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.5635792778649922, |
|
"grad_norm": 0.2999591827392578, |
|
"learning_rate": 4.372056514913658e-06, |
|
"loss": 1.0542, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.5643642072213501, |
|
"grad_norm": 0.2673451006412506, |
|
"learning_rate": 4.364207221350079e-06, |
|
"loss": 1.0464, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.565149136577708, |
|
"grad_norm": 0.27569347620010376, |
|
"learning_rate": 4.3563579277865e-06, |
|
"loss": 1.1031, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.5659340659340659, |
|
"grad_norm": 0.30383068323135376, |
|
"learning_rate": 4.34850863422292e-06, |
|
"loss": 1.145, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.5667189952904239, |
|
"grad_norm": 0.2751275300979614, |
|
"learning_rate": 4.340659340659341e-06, |
|
"loss": 1.0859, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.5675039246467818, |
|
"grad_norm": 0.27169832587242126, |
|
"learning_rate": 4.332810047095762e-06, |
|
"loss": 1.0594, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.5682888540031397, |
|
"grad_norm": 0.28917625546455383, |
|
"learning_rate": 4.324960753532183e-06, |
|
"loss": 1.0823, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.5690737833594977, |
|
"grad_norm": 0.2915303409099579, |
|
"learning_rate": 4.317111459968603e-06, |
|
"loss": 1.0867, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.5698587127158555, |
|
"grad_norm": 0.2857419550418854, |
|
"learning_rate": 4.309262166405024e-06, |
|
"loss": 1.0843, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.5706436420722135, |
|
"grad_norm": 0.2684113085269928, |
|
"learning_rate": 4.301412872841444e-06, |
|
"loss": 1.0595, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.5714285714285714, |
|
"grad_norm": 0.2712114155292511, |
|
"learning_rate": 4.293563579277865e-06, |
|
"loss": 1.0989, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.5722135007849294, |
|
"grad_norm": 0.2832731604576111, |
|
"learning_rate": 4.2857142857142855e-06, |
|
"loss": 1.0498, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.5729984301412873, |
|
"grad_norm": 0.44053250551223755, |
|
"learning_rate": 4.277864992150707e-06, |
|
"loss": 1.0979, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.5737833594976453, |
|
"grad_norm": 0.27188801765441895, |
|
"learning_rate": 4.270015698587127e-06, |
|
"loss": 1.0941, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.5745682888540031, |
|
"grad_norm": 0.28184184432029724, |
|
"learning_rate": 4.262166405023548e-06, |
|
"loss": 1.1324, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.5753532182103611, |
|
"grad_norm": 0.27184492349624634, |
|
"learning_rate": 4.254317111459969e-06, |
|
"loss": 1.0883, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.576138147566719, |
|
"grad_norm": 0.2704511284828186, |
|
"learning_rate": 4.24646781789639e-06, |
|
"loss": 1.0746, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.5769230769230769, |
|
"grad_norm": 0.2777252793312073, |
|
"learning_rate": 4.2386185243328105e-06, |
|
"loss": 1.0202, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.5777080062794349, |
|
"grad_norm": 0.279325932264328, |
|
"learning_rate": 4.230769230769231e-06, |
|
"loss": 1.0334, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.5784929356357927, |
|
"grad_norm": 0.3658686876296997, |
|
"learning_rate": 4.222919937205652e-06, |
|
"loss": 1.0314, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.5792778649921507, |
|
"grad_norm": 0.2697162926197052, |
|
"learning_rate": 4.215070643642072e-06, |
|
"loss": 1.0703, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.5800627943485086, |
|
"grad_norm": 0.27418413758277893, |
|
"learning_rate": 4.207221350078493e-06, |
|
"loss": 1.0345, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.5808477237048666, |
|
"grad_norm": 0.2849864065647125, |
|
"learning_rate": 4.199372056514914e-06, |
|
"loss": 1.0609, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.5816326530612245, |
|
"grad_norm": 0.29697927832603455, |
|
"learning_rate": 4.191522762951335e-06, |
|
"loss": 1.0578, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.5824175824175825, |
|
"grad_norm": 0.2827671766281128, |
|
"learning_rate": 4.183673469387755e-06, |
|
"loss": 1.0264, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.5832025117739403, |
|
"grad_norm": 0.26230207085609436, |
|
"learning_rate": 4.175824175824177e-06, |
|
"loss": 1.0614, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.5839874411302983, |
|
"grad_norm": 0.28049588203430176, |
|
"learning_rate": 4.167974882260597e-06, |
|
"loss": 1.0784, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.5847723704866562, |
|
"grad_norm": 0.2954070270061493, |
|
"learning_rate": 4.160125588697018e-06, |
|
"loss": 1.1016, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.5855572998430141, |
|
"grad_norm": 0.3010290861129761, |
|
"learning_rate": 4.152276295133438e-06, |
|
"loss": 1.0409, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.5863422291993721, |
|
"grad_norm": 0.26996880769729614, |
|
"learning_rate": 4.144427001569859e-06, |
|
"loss": 1.0828, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.5871271585557299, |
|
"grad_norm": 0.27959656715393066, |
|
"learning_rate": 4.1365777080062794e-06, |
|
"loss": 0.9985, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.5879120879120879, |
|
"grad_norm": 0.26965901255607605, |
|
"learning_rate": 4.1287284144427e-06, |
|
"loss": 1.0649, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.5886970172684458, |
|
"grad_norm": 0.2751823365688324, |
|
"learning_rate": 4.120879120879121e-06, |
|
"loss": 1.0422, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.5894819466248038, |
|
"grad_norm": 0.27731597423553467, |
|
"learning_rate": 4.113029827315542e-06, |
|
"loss": 1.0927, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.5902668759811617, |
|
"grad_norm": 0.2711530029773712, |
|
"learning_rate": 4.1051805337519625e-06, |
|
"loss": 1.055, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.5910518053375197, |
|
"grad_norm": 0.28580254316329956, |
|
"learning_rate": 4.097331240188384e-06, |
|
"loss": 1.0859, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.5918367346938775, |
|
"grad_norm": 0.2740454077720642, |
|
"learning_rate": 4.0894819466248045e-06, |
|
"loss": 1.0285, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.5926216640502355, |
|
"grad_norm": 0.28732219338417053, |
|
"learning_rate": 4.081632653061225e-06, |
|
"loss": 1.0953, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.5934065934065934, |
|
"grad_norm": 0.2803926467895508, |
|
"learning_rate": 4.0737833594976456e-06, |
|
"loss": 1.0135, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.5941915227629513, |
|
"grad_norm": 0.2755890488624573, |
|
"learning_rate": 4.065934065934066e-06, |
|
"loss": 1.0422, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.5949764521193093, |
|
"grad_norm": 0.28653979301452637, |
|
"learning_rate": 4.058084772370487e-06, |
|
"loss": 1.0301, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.5957613814756672, |
|
"grad_norm": 0.2685067355632782, |
|
"learning_rate": 4.050235478806907e-06, |
|
"loss": 1.03, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.5965463108320251, |
|
"grad_norm": 0.2731209397315979, |
|
"learning_rate": 4.042386185243329e-06, |
|
"loss": 1.0346, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.597331240188383, |
|
"grad_norm": 0.29851043224334717, |
|
"learning_rate": 4.034536891679749e-06, |
|
"loss": 1.0961, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.598116169544741, |
|
"grad_norm": 0.28820693492889404, |
|
"learning_rate": 4.02668759811617e-06, |
|
"loss": 1.0942, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.5989010989010989, |
|
"grad_norm": 0.28037169575691223, |
|
"learning_rate": 4.018838304552591e-06, |
|
"loss": 1.0647, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.5996860282574569, |
|
"grad_norm": 0.27676883339881897, |
|
"learning_rate": 4.010989010989012e-06, |
|
"loss": 1.0605, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.6004709576138147, |
|
"grad_norm": 0.274884432554245, |
|
"learning_rate": 4.003139717425432e-06, |
|
"loss": 1.0577, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.6012558869701727, |
|
"grad_norm": 0.2712654173374176, |
|
"learning_rate": 3.995290423861853e-06, |
|
"loss": 1.0871, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.6020408163265306, |
|
"grad_norm": 0.2694210708141327, |
|
"learning_rate": 3.987441130298273e-06, |
|
"loss": 1.0412, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.6028257456828885, |
|
"grad_norm": 0.2664571702480316, |
|
"learning_rate": 3.979591836734694e-06, |
|
"loss": 1.0413, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.6036106750392465, |
|
"grad_norm": 0.2823816239833832, |
|
"learning_rate": 3.9717425431711145e-06, |
|
"loss": 1.0511, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.6043956043956044, |
|
"grad_norm": 0.27906733751296997, |
|
"learning_rate": 3.963893249607536e-06, |
|
"loss": 1.059, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.6051805337519623, |
|
"grad_norm": 0.26882991194725037, |
|
"learning_rate": 3.9560439560439565e-06, |
|
"loss": 1.0445, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.6059654631083202, |
|
"grad_norm": 0.273629754781723, |
|
"learning_rate": 3.948194662480377e-06, |
|
"loss": 1.0566, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.6067503924646782, |
|
"grad_norm": 0.28593695163726807, |
|
"learning_rate": 3.940345368916798e-06, |
|
"loss": 1.0857, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.6075353218210361, |
|
"grad_norm": 0.2721453309059143, |
|
"learning_rate": 3.932496075353219e-06, |
|
"loss": 1.0474, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.6083202511773941, |
|
"grad_norm": 0.2677747905254364, |
|
"learning_rate": 3.9246467817896395e-06, |
|
"loss": 1.0293, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.609105180533752, |
|
"grad_norm": 0.27768194675445557, |
|
"learning_rate": 3.91679748822606e-06, |
|
"loss": 1.0645, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.6098901098901099, |
|
"grad_norm": 0.2910935878753662, |
|
"learning_rate": 3.908948194662481e-06, |
|
"loss": 1.122, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.6106750392464678, |
|
"grad_norm": 0.2711617946624756, |
|
"learning_rate": 3.901098901098901e-06, |
|
"loss": 1.0663, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.6114599686028257, |
|
"grad_norm": 0.2801941931247711, |
|
"learning_rate": 3.893249607535322e-06, |
|
"loss": 1.057, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.6122448979591837, |
|
"grad_norm": 0.2923711836338043, |
|
"learning_rate": 3.885400313971743e-06, |
|
"loss": 1.1437, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.6130298273155416, |
|
"grad_norm": 0.2791869342327118, |
|
"learning_rate": 3.877551020408164e-06, |
|
"loss": 1.0828, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.6138147566718996, |
|
"grad_norm": 0.34363803267478943, |
|
"learning_rate": 3.869701726844584e-06, |
|
"loss": 1.0993, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.6145996860282574, |
|
"grad_norm": 0.2742525041103363, |
|
"learning_rate": 3.861852433281005e-06, |
|
"loss": 1.0626, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.6153846153846154, |
|
"grad_norm": 0.2778567671775818, |
|
"learning_rate": 3.854003139717426e-06, |
|
"loss": 1.0768, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.6161695447409733, |
|
"grad_norm": 0.274810791015625, |
|
"learning_rate": 3.846153846153847e-06, |
|
"loss": 1.07, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.6169544740973313, |
|
"grad_norm": 0.33274561166763306, |
|
"learning_rate": 3.838304552590267e-06, |
|
"loss": 1.0368, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.6177394034536892, |
|
"grad_norm": 0.2886803448200226, |
|
"learning_rate": 3.830455259026688e-06, |
|
"loss": 1.0458, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.6185243328100472, |
|
"grad_norm": 0.3283863961696625, |
|
"learning_rate": 3.8226059654631085e-06, |
|
"loss": 1.0733, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.619309262166405, |
|
"grad_norm": 0.30149760842323303, |
|
"learning_rate": 3.814756671899529e-06, |
|
"loss": 1.09, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.6200941915227629, |
|
"grad_norm": 0.2996380627155304, |
|
"learning_rate": 3.80690737833595e-06, |
|
"loss": 1.0629, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.6208791208791209, |
|
"grad_norm": 0.3076927661895752, |
|
"learning_rate": 3.7990580847723706e-06, |
|
"loss": 1.051, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.6216640502354788, |
|
"grad_norm": 0.27474308013916016, |
|
"learning_rate": 3.7912087912087915e-06, |
|
"loss": 1.0259, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.6224489795918368, |
|
"grad_norm": 0.2838291823863983, |
|
"learning_rate": 3.783359497645212e-06, |
|
"loss": 1.0489, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.6232339089481946, |
|
"grad_norm": 0.31608259677886963, |
|
"learning_rate": 3.7755102040816327e-06, |
|
"loss": 1.0994, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.6240188383045526, |
|
"grad_norm": 0.29825273156166077, |
|
"learning_rate": 3.767660910518054e-06, |
|
"loss": 1.0947, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.6248037676609105, |
|
"grad_norm": 0.28364327549934387, |
|
"learning_rate": 3.7598116169544746e-06, |
|
"loss": 1.0869, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.6255886970172685, |
|
"grad_norm": 0.2713553011417389, |
|
"learning_rate": 3.751962323390895e-06, |
|
"loss": 1.0656, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.6263736263736264, |
|
"grad_norm": 0.27098798751831055, |
|
"learning_rate": 3.744113029827316e-06, |
|
"loss": 1.073, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.6271585557299842, |
|
"grad_norm": 0.28046950697898865, |
|
"learning_rate": 3.7362637362637367e-06, |
|
"loss": 1.0539, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.6279434850863422, |
|
"grad_norm": 0.2658367156982422, |
|
"learning_rate": 3.7284144427001573e-06, |
|
"loss": 1.0596, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.6287284144427001, |
|
"grad_norm": 0.2806791365146637, |
|
"learning_rate": 3.720565149136578e-06, |
|
"loss": 1.0867, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.6295133437990581, |
|
"grad_norm": 0.27192607522010803, |
|
"learning_rate": 3.712715855572999e-06, |
|
"loss": 1.0323, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.630298273155416, |
|
"grad_norm": 0.29174211621284485, |
|
"learning_rate": 3.7048665620094194e-06, |
|
"loss": 1.0436, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.631083202511774, |
|
"grad_norm": 0.2949182093143463, |
|
"learning_rate": 3.69701726844584e-06, |
|
"loss": 1.0729, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.6318681318681318, |
|
"grad_norm": 0.28802135586738586, |
|
"learning_rate": 3.6891679748822605e-06, |
|
"loss": 1.0826, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.6326530612244898, |
|
"grad_norm": 0.27609679102897644, |
|
"learning_rate": 3.681318681318682e-06, |
|
"loss": 1.0617, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.6334379905808477, |
|
"grad_norm": 0.30500441789627075, |
|
"learning_rate": 3.6734693877551024e-06, |
|
"loss": 1.0943, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.6342229199372057, |
|
"grad_norm": 0.2916868329048157, |
|
"learning_rate": 3.6656200941915234e-06, |
|
"loss": 1.1224, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.6350078492935636, |
|
"grad_norm": 0.3102125823497772, |
|
"learning_rate": 3.657770800627944e-06, |
|
"loss": 1.1176, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.6357927786499215, |
|
"grad_norm": 0.2802576422691345, |
|
"learning_rate": 3.6499215070643645e-06, |
|
"loss": 1.0794, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.6365777080062794, |
|
"grad_norm": 0.3083432912826538, |
|
"learning_rate": 3.642072213500785e-06, |
|
"loss": 1.1249, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.6373626373626373, |
|
"grad_norm": 0.30711638927459717, |
|
"learning_rate": 3.634222919937206e-06, |
|
"loss": 1.0287, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.6381475667189953, |
|
"grad_norm": 0.3342186510562897, |
|
"learning_rate": 3.6263736263736266e-06, |
|
"loss": 1.0822, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.6389324960753532, |
|
"grad_norm": 0.27452296018600464, |
|
"learning_rate": 3.618524332810047e-06, |
|
"loss": 1.0322, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.6397174254317112, |
|
"grad_norm": 0.2835961580276489, |
|
"learning_rate": 3.6106750392464677e-06, |
|
"loss": 1.0897, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.640502354788069, |
|
"grad_norm": 0.27237561345100403, |
|
"learning_rate": 3.6028257456828887e-06, |
|
"loss": 1.0461, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.641287284144427, |
|
"grad_norm": 0.31648552417755127, |
|
"learning_rate": 3.5949764521193097e-06, |
|
"loss": 1.0393, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.6420722135007849, |
|
"grad_norm": 0.27146708965301514, |
|
"learning_rate": 3.5871271585557307e-06, |
|
"loss": 1.0706, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.6428571428571429, |
|
"grad_norm": 0.2728872299194336, |
|
"learning_rate": 3.5792778649921512e-06, |
|
"loss": 1.0361, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.6436420722135008, |
|
"grad_norm": 0.2697795033454895, |
|
"learning_rate": 3.5714285714285718e-06, |
|
"loss": 1.0238, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.6444270015698587, |
|
"grad_norm": 0.3166142702102661, |
|
"learning_rate": 3.5635792778649923e-06, |
|
"loss": 1.1152, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.6452119309262166, |
|
"grad_norm": 0.3062928020954132, |
|
"learning_rate": 3.5557299843014133e-06, |
|
"loss": 1.1208, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.6459968602825745, |
|
"grad_norm": 0.3098381757736206, |
|
"learning_rate": 3.547880690737834e-06, |
|
"loss": 1.1239, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.6467817896389325, |
|
"grad_norm": 0.2634499669075012, |
|
"learning_rate": 3.5400313971742544e-06, |
|
"loss": 1.0154, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.6475667189952904, |
|
"grad_norm": 0.27799683809280396, |
|
"learning_rate": 3.532182103610675e-06, |
|
"loss": 1.0574, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.6483516483516484, |
|
"grad_norm": 0.29064470529556274, |
|
"learning_rate": 3.524332810047096e-06, |
|
"loss": 1.0962, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.6491365777080063, |
|
"grad_norm": 0.3558158874511719, |
|
"learning_rate": 3.516483516483517e-06, |
|
"loss": 1.1029, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.6499215070643642, |
|
"grad_norm": 0.3792993426322937, |
|
"learning_rate": 3.5086342229199375e-06, |
|
"loss": 1.0777, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.6507064364207221, |
|
"grad_norm": 0.29269522428512573, |
|
"learning_rate": 3.5007849293563585e-06, |
|
"loss": 1.0245, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.6514913657770801, |
|
"grad_norm": 0.28923287987709045, |
|
"learning_rate": 3.492935635792779e-06, |
|
"loss": 1.0701, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.652276295133438, |
|
"grad_norm": 0.285043865442276, |
|
"learning_rate": 3.4850863422291996e-06, |
|
"loss": 1.0538, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.6530612244897959, |
|
"grad_norm": 0.2893431782722473, |
|
"learning_rate": 3.4772370486656206e-06, |
|
"loss": 1.038, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.6538461538461539, |
|
"grad_norm": 0.28342151641845703, |
|
"learning_rate": 3.469387755102041e-06, |
|
"loss": 1.0341, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.6546310832025117, |
|
"grad_norm": 0.27439647912979126, |
|
"learning_rate": 3.4615384615384617e-06, |
|
"loss": 1.0807, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.6554160125588697, |
|
"grad_norm": 0.283348023891449, |
|
"learning_rate": 3.4536891679748822e-06, |
|
"loss": 1.0321, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.6562009419152276, |
|
"grad_norm": 0.3052699565887451, |
|
"learning_rate": 3.4458398744113032e-06, |
|
"loss": 1.0977, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.6569858712715856, |
|
"grad_norm": 0.30584266781806946, |
|
"learning_rate": 3.4379905808477238e-06, |
|
"loss": 1.0312, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.6577708006279435, |
|
"grad_norm": 0.2961052656173706, |
|
"learning_rate": 3.4301412872841448e-06, |
|
"loss": 1.064, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.6585557299843015, |
|
"grad_norm": 0.32309481501579285, |
|
"learning_rate": 3.4222919937205657e-06, |
|
"loss": 1.087, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.6593406593406593, |
|
"grad_norm": 0.2820388674736023, |
|
"learning_rate": 3.4144427001569863e-06, |
|
"loss": 1.062, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.6601255886970173, |
|
"grad_norm": 0.31144237518310547, |
|
"learning_rate": 3.406593406593407e-06, |
|
"loss": 1.0137, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.6609105180533752, |
|
"grad_norm": 0.2903454303741455, |
|
"learning_rate": 3.398744113029828e-06, |
|
"loss": 1.0571, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.6616954474097331, |
|
"grad_norm": 0.29997384548187256, |
|
"learning_rate": 3.3908948194662484e-06, |
|
"loss": 1.0569, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.6624803767660911, |
|
"grad_norm": 0.34617769718170166, |
|
"learning_rate": 3.383045525902669e-06, |
|
"loss": 1.0925, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.6632653061224489, |
|
"grad_norm": 0.2661650776863098, |
|
"learning_rate": 3.3751962323390895e-06, |
|
"loss": 1.0299, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.6640502354788069, |
|
"grad_norm": 0.2766907215118408, |
|
"learning_rate": 3.3673469387755105e-06, |
|
"loss": 1.0807, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.6648351648351648, |
|
"grad_norm": 0.2823966145515442, |
|
"learning_rate": 3.359497645211931e-06, |
|
"loss": 1.0796, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.6656200941915228, |
|
"grad_norm": 0.32514306902885437, |
|
"learning_rate": 3.3516483516483516e-06, |
|
"loss": 1.1211, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.6664050235478807, |
|
"grad_norm": 0.3008269965648651, |
|
"learning_rate": 3.343799058084773e-06, |
|
"loss": 1.0751, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.6671899529042387, |
|
"grad_norm": 0.30757200717926025, |
|
"learning_rate": 3.3359497645211936e-06, |
|
"loss": 1.1073, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.6679748822605965, |
|
"grad_norm": 0.2902880609035492, |
|
"learning_rate": 3.328100470957614e-06, |
|
"loss": 1.0854, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.6687598116169545, |
|
"grad_norm": 0.2838514447212219, |
|
"learning_rate": 3.320251177394035e-06, |
|
"loss": 1.0474, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.6695447409733124, |
|
"grad_norm": 0.30020883679389954, |
|
"learning_rate": 3.3124018838304557e-06, |
|
"loss": 1.0338, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.6703296703296703, |
|
"grad_norm": 0.29149070382118225, |
|
"learning_rate": 3.304552590266876e-06, |
|
"loss": 1.0447, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.6711145996860283, |
|
"grad_norm": 0.2783101201057434, |
|
"learning_rate": 3.2967032967032968e-06, |
|
"loss": 1.0215, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.6718995290423861, |
|
"grad_norm": 0.2824500501155853, |
|
"learning_rate": 3.2888540031397177e-06, |
|
"loss": 1.002, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.6726844583987441, |
|
"grad_norm": 0.2815590798854828, |
|
"learning_rate": 3.2810047095761383e-06, |
|
"loss": 1.1188, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.673469387755102, |
|
"grad_norm": 0.2877782881259918, |
|
"learning_rate": 3.273155416012559e-06, |
|
"loss": 1.0294, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.67425431711146, |
|
"grad_norm": 0.27774369716644287, |
|
"learning_rate": 3.2653061224489794e-06, |
|
"loss": 1.0293, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.6750392464678179, |
|
"grad_norm": 0.3431270122528076, |
|
"learning_rate": 3.257456828885401e-06, |
|
"loss": 1.072, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.6758241758241759, |
|
"grad_norm": 0.3327620029449463, |
|
"learning_rate": 3.2496075353218214e-06, |
|
"loss": 1.095, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.6766091051805337, |
|
"grad_norm": 0.288352370262146, |
|
"learning_rate": 3.2417582417582424e-06, |
|
"loss": 1.0711, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.6773940345368917, |
|
"grad_norm": 0.2850242257118225, |
|
"learning_rate": 3.233908948194663e-06, |
|
"loss": 1.0785, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.6781789638932496, |
|
"grad_norm": 0.2831905782222748, |
|
"learning_rate": 3.2260596546310835e-06, |
|
"loss": 1.0341, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.6789638932496075, |
|
"grad_norm": 0.284157931804657, |
|
"learning_rate": 3.218210361067504e-06, |
|
"loss": 1.0639, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.6797488226059655, |
|
"grad_norm": 0.27813270688056946, |
|
"learning_rate": 3.210361067503925e-06, |
|
"loss": 1.0721, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.6805337519623234, |
|
"grad_norm": 0.28956329822540283, |
|
"learning_rate": 3.2025117739403456e-06, |
|
"loss": 1.0826, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.6813186813186813, |
|
"grad_norm": 0.29287075996398926, |
|
"learning_rate": 3.194662480376766e-06, |
|
"loss": 1.1073, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.6821036106750392, |
|
"grad_norm": 0.2984355092048645, |
|
"learning_rate": 3.1868131868131867e-06, |
|
"loss": 1.0131, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.6828885400313972, |
|
"grad_norm": 0.2821354269981384, |
|
"learning_rate": 3.1789638932496077e-06, |
|
"loss": 1.0452, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.6836734693877551, |
|
"grad_norm": 0.27414095401763916, |
|
"learning_rate": 3.1711145996860286e-06, |
|
"loss": 1.0228, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.6844583987441131, |
|
"grad_norm": 0.2787257730960846, |
|
"learning_rate": 3.1632653061224496e-06, |
|
"loss": 1.0317, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.685243328100471, |
|
"grad_norm": 0.2801453769207001, |
|
"learning_rate": 3.15541601255887e-06, |
|
"loss": 1.0305, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.6860282574568289, |
|
"grad_norm": 0.29602035880088806, |
|
"learning_rate": 3.1475667189952907e-06, |
|
"loss": 1.0955, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.6868131868131868, |
|
"grad_norm": 0.301297664642334, |
|
"learning_rate": 3.1397174254317113e-06, |
|
"loss": 1.036, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.6875981161695447, |
|
"grad_norm": 0.3049217462539673, |
|
"learning_rate": 3.1318681318681323e-06, |
|
"loss": 1.1071, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.6883830455259027, |
|
"grad_norm": 0.2837543785572052, |
|
"learning_rate": 3.124018838304553e-06, |
|
"loss": 1.0273, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.6891679748822606, |
|
"grad_norm": 0.28585192561149597, |
|
"learning_rate": 3.1161695447409734e-06, |
|
"loss": 1.0464, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.6899529042386185, |
|
"grad_norm": 0.2744940519332886, |
|
"learning_rate": 3.108320251177394e-06, |
|
"loss": 1.0626, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.6907378335949764, |
|
"grad_norm": 0.29953551292419434, |
|
"learning_rate": 3.100470957613815e-06, |
|
"loss": 1.0375, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.6915227629513344, |
|
"grad_norm": 0.30678224563598633, |
|
"learning_rate": 3.092621664050236e-06, |
|
"loss": 1.0765, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.6923076923076923, |
|
"grad_norm": 0.29770031571388245, |
|
"learning_rate": 3.084772370486657e-06, |
|
"loss": 1.0751, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.6930926216640503, |
|
"grad_norm": 0.2807864546775818, |
|
"learning_rate": 3.0769230769230774e-06, |
|
"loss": 1.0866, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.6938775510204082, |
|
"grad_norm": 0.29043442010879517, |
|
"learning_rate": 3.069073783359498e-06, |
|
"loss": 1.0886, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.6946624803767661, |
|
"grad_norm": 0.31291401386260986, |
|
"learning_rate": 3.0612244897959185e-06, |
|
"loss": 1.0481, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.695447409733124, |
|
"grad_norm": 0.28840577602386475, |
|
"learning_rate": 3.0533751962323395e-06, |
|
"loss": 1.0756, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.6962323390894819, |
|
"grad_norm": 0.30978354811668396, |
|
"learning_rate": 3.04552590266876e-06, |
|
"loss": 1.0258, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.6970172684458399, |
|
"grad_norm": 0.29402005672454834, |
|
"learning_rate": 3.0376766091051806e-06, |
|
"loss": 1.0499, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.6978021978021978, |
|
"grad_norm": 0.29459160566329956, |
|
"learning_rate": 3.029827315541601e-06, |
|
"loss": 1.1056, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.6985871271585558, |
|
"grad_norm": 0.3101595938205719, |
|
"learning_rate": 3.021978021978022e-06, |
|
"loss": 1.145, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.6993720565149136, |
|
"grad_norm": 0.2799089550971985, |
|
"learning_rate": 3.0141287284144427e-06, |
|
"loss": 1.0958, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.7001569858712716, |
|
"grad_norm": 0.29476428031921387, |
|
"learning_rate": 3.006279434850864e-06, |
|
"loss": 1.109, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.7009419152276295, |
|
"grad_norm": 0.291610985994339, |
|
"learning_rate": 2.9984301412872847e-06, |
|
"loss": 1.052, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.7017268445839875, |
|
"grad_norm": 0.29794690012931824, |
|
"learning_rate": 2.9905808477237053e-06, |
|
"loss": 1.083, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.7025117739403454, |
|
"grad_norm": 0.28312987089157104, |
|
"learning_rate": 2.982731554160126e-06, |
|
"loss": 1.0448, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.7032967032967034, |
|
"grad_norm": 0.28827813267707825, |
|
"learning_rate": 2.974882260596547e-06, |
|
"loss": 1.0442, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.7040816326530612, |
|
"grad_norm": 0.29806190729141235, |
|
"learning_rate": 2.9670329670329673e-06, |
|
"loss": 1.0796, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.7048665620094191, |
|
"grad_norm": 0.2928798496723175, |
|
"learning_rate": 2.959183673469388e-06, |
|
"loss": 1.0792, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.7056514913657771, |
|
"grad_norm": 0.2975620925426483, |
|
"learning_rate": 2.9513343799058085e-06, |
|
"loss": 1.0481, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.706436420722135, |
|
"grad_norm": 0.2881799042224884, |
|
"learning_rate": 2.9434850863422294e-06, |
|
"loss": 1.0588, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.707221350078493, |
|
"grad_norm": 0.289421021938324, |
|
"learning_rate": 2.93563579277865e-06, |
|
"loss": 1.1009, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.7080062794348508, |
|
"grad_norm": 0.30106186866760254, |
|
"learning_rate": 2.9277864992150706e-06, |
|
"loss": 1.1007, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.7087912087912088, |
|
"grad_norm": 0.2982400059700012, |
|
"learning_rate": 2.919937205651492e-06, |
|
"loss": 1.0781, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.7095761381475667, |
|
"grad_norm": 0.27340105175971985, |
|
"learning_rate": 2.9120879120879125e-06, |
|
"loss": 1.0401, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.7103610675039247, |
|
"grad_norm": 0.2898809313774109, |
|
"learning_rate": 2.904238618524333e-06, |
|
"loss": 1.1035, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.7111459968602826, |
|
"grad_norm": 0.28713247179985046, |
|
"learning_rate": 2.896389324960754e-06, |
|
"loss": 1.043, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.7119309262166404, |
|
"grad_norm": 0.2877185046672821, |
|
"learning_rate": 2.8885400313971746e-06, |
|
"loss": 1.0845, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.7127158555729984, |
|
"grad_norm": 0.2887587249279022, |
|
"learning_rate": 2.880690737833595e-06, |
|
"loss": 1.0751, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.7135007849293563, |
|
"grad_norm": 0.2877505123615265, |
|
"learning_rate": 2.8728414442700157e-06, |
|
"loss": 1.0312, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.7142857142857143, |
|
"grad_norm": 0.28425735235214233, |
|
"learning_rate": 2.8649921507064367e-06, |
|
"loss": 1.0554, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.7150706436420722, |
|
"grad_norm": 0.2878026068210602, |
|
"learning_rate": 2.8571428571428573e-06, |
|
"loss": 1.0041, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.7158555729984302, |
|
"grad_norm": 0.4087301194667816, |
|
"learning_rate": 2.849293563579278e-06, |
|
"loss": 1.0523, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.716640502354788, |
|
"grad_norm": 0.2949962913990021, |
|
"learning_rate": 2.8414442700156984e-06, |
|
"loss": 1.0276, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.717425431711146, |
|
"grad_norm": 0.2939402163028717, |
|
"learning_rate": 2.8335949764521198e-06, |
|
"loss": 1.0428, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.7182103610675039, |
|
"grad_norm": 0.27334320545196533, |
|
"learning_rate": 2.8257456828885403e-06, |
|
"loss": 1.0218, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.7189952904238619, |
|
"grad_norm": 0.28242912888526917, |
|
"learning_rate": 2.8178963893249613e-06, |
|
"loss": 1.0306, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.7197802197802198, |
|
"grad_norm": 0.30455905199050903, |
|
"learning_rate": 2.810047095761382e-06, |
|
"loss": 1.0888, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 0.7205651491365777, |
|
"grad_norm": 0.31174805760383606, |
|
"learning_rate": 2.8021978021978024e-06, |
|
"loss": 1.041, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.7213500784929356, |
|
"grad_norm": 0.30026131868362427, |
|
"learning_rate": 2.794348508634223e-06, |
|
"loss": 1.0294, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 0.7221350078492935, |
|
"grad_norm": 0.2858772575855255, |
|
"learning_rate": 2.786499215070644e-06, |
|
"loss": 1.0676, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.7229199372056515, |
|
"grad_norm": 0.3418976962566376, |
|
"learning_rate": 2.7786499215070645e-06, |
|
"loss": 1.0551, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.7237048665620094, |
|
"grad_norm": 0.31597593426704407, |
|
"learning_rate": 2.770800627943485e-06, |
|
"loss": 1.063, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 0.7244897959183674, |
|
"grad_norm": 0.2929779291152954, |
|
"learning_rate": 2.7629513343799056e-06, |
|
"loss": 1.0147, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 0.7252747252747253, |
|
"grad_norm": 0.2997245192527771, |
|
"learning_rate": 2.7551020408163266e-06, |
|
"loss": 1.1207, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.7260596546310832, |
|
"grad_norm": 0.2890755832195282, |
|
"learning_rate": 2.7472527472527476e-06, |
|
"loss": 1.0533, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.7268445839874411, |
|
"grad_norm": 0.29468992352485657, |
|
"learning_rate": 2.7394034536891686e-06, |
|
"loss": 1.0764, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 0.7276295133437991, |
|
"grad_norm": 0.28897175192832947, |
|
"learning_rate": 2.731554160125589e-06, |
|
"loss": 1.0755, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.728414442700157, |
|
"grad_norm": 0.30992481112480164, |
|
"learning_rate": 2.7237048665620097e-06, |
|
"loss": 1.0074, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 0.7291993720565149, |
|
"grad_norm": 0.29595842957496643, |
|
"learning_rate": 2.7158555729984302e-06, |
|
"loss": 1.1159, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 0.7299843014128728, |
|
"grad_norm": 0.2836659550666809, |
|
"learning_rate": 2.7080062794348512e-06, |
|
"loss": 1.0807, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.7307692307692307, |
|
"grad_norm": 0.2799653708934784, |
|
"learning_rate": 2.7001569858712718e-06, |
|
"loss": 1.0111, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 0.7315541601255887, |
|
"grad_norm": 0.29231712222099304, |
|
"learning_rate": 2.6923076923076923e-06, |
|
"loss": 1.039, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 0.7323390894819466, |
|
"grad_norm": 0.2751501202583313, |
|
"learning_rate": 2.684458398744113e-06, |
|
"loss": 1.0743, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.7331240188383046, |
|
"grad_norm": 0.2840130031108856, |
|
"learning_rate": 2.676609105180534e-06, |
|
"loss": 1.0533, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 0.7339089481946625, |
|
"grad_norm": 0.28328344225883484, |
|
"learning_rate": 2.668759811616955e-06, |
|
"loss": 1.0726, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.7346938775510204, |
|
"grad_norm": 0.2846105694770813, |
|
"learning_rate": 2.660910518053376e-06, |
|
"loss": 1.0482, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.7354788069073783, |
|
"grad_norm": 0.2846198081970215, |
|
"learning_rate": 2.6530612244897964e-06, |
|
"loss": 1.0549, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 0.7362637362637363, |
|
"grad_norm": 0.2910875678062439, |
|
"learning_rate": 2.645211930926217e-06, |
|
"loss": 1.0597, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 0.7370486656200942, |
|
"grad_norm": 0.3099419176578522, |
|
"learning_rate": 2.6373626373626375e-06, |
|
"loss": 1.1132, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.7378335949764521, |
|
"grad_norm": 0.30356669425964355, |
|
"learning_rate": 2.6295133437990585e-06, |
|
"loss": 1.0268, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.7386185243328101, |
|
"grad_norm": 0.28936824202537537, |
|
"learning_rate": 2.621664050235479e-06, |
|
"loss": 1.0237, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 0.7394034536891679, |
|
"grad_norm": 0.2909795641899109, |
|
"learning_rate": 2.6138147566718996e-06, |
|
"loss": 1.0425, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 0.7401883830455259, |
|
"grad_norm": 0.29834648966789246, |
|
"learning_rate": 2.60596546310832e-06, |
|
"loss": 1.0731, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 0.7409733124018838, |
|
"grad_norm": 0.3154754042625427, |
|
"learning_rate": 2.598116169544741e-06, |
|
"loss": 1.144, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 0.7417582417582418, |
|
"grad_norm": 0.2903672456741333, |
|
"learning_rate": 2.5902668759811617e-06, |
|
"loss": 1.0753, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.7425431711145997, |
|
"grad_norm": 0.28852578997612, |
|
"learning_rate": 2.582417582417583e-06, |
|
"loss": 1.0292, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 0.7433281004709577, |
|
"grad_norm": 0.2857038080692291, |
|
"learning_rate": 2.5745682888540036e-06, |
|
"loss": 1.0717, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 0.7441130298273155, |
|
"grad_norm": 0.2909829914569855, |
|
"learning_rate": 2.566718995290424e-06, |
|
"loss": 1.0209, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 0.7448979591836735, |
|
"grad_norm": 0.2876448631286621, |
|
"learning_rate": 2.5588697017268448e-06, |
|
"loss": 1.0688, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 0.7456828885400314, |
|
"grad_norm": 0.2869911789894104, |
|
"learning_rate": 2.5510204081632657e-06, |
|
"loss": 1.0332, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.7464678178963893, |
|
"grad_norm": 0.2981649935245514, |
|
"learning_rate": 2.5431711145996863e-06, |
|
"loss": 1.0635, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 0.7472527472527473, |
|
"grad_norm": 0.28783732652664185, |
|
"learning_rate": 2.535321821036107e-06, |
|
"loss": 1.0492, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 0.7480376766091051, |
|
"grad_norm": 0.28739696741104126, |
|
"learning_rate": 2.5274725274725274e-06, |
|
"loss": 1.0318, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 0.7488226059654631, |
|
"grad_norm": 0.28677845001220703, |
|
"learning_rate": 2.5196232339089484e-06, |
|
"loss": 1.032, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 0.749607535321821, |
|
"grad_norm": 0.3001886010169983, |
|
"learning_rate": 2.511773940345369e-06, |
|
"loss": 1.0855, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.750392464678179, |
|
"grad_norm": 0.2894863486289978, |
|
"learning_rate": 2.5039246467817895e-06, |
|
"loss": 1.044, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 0.7511773940345369, |
|
"grad_norm": 0.29826030135154724, |
|
"learning_rate": 2.4960753532182105e-06, |
|
"loss": 1.0313, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 0.7519623233908949, |
|
"grad_norm": 0.30050304532051086, |
|
"learning_rate": 2.488226059654631e-06, |
|
"loss": 1.0575, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.7527472527472527, |
|
"grad_norm": 0.3099324107170105, |
|
"learning_rate": 2.480376766091052e-06, |
|
"loss": 1.0636, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 0.7535321821036107, |
|
"grad_norm": 0.2929956912994385, |
|
"learning_rate": 2.472527472527473e-06, |
|
"loss": 1.0516, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.7543171114599686, |
|
"grad_norm": 0.29828134179115295, |
|
"learning_rate": 2.4646781789638936e-06, |
|
"loss": 1.0582, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 0.7551020408163265, |
|
"grad_norm": 0.3314920663833618, |
|
"learning_rate": 2.456828885400314e-06, |
|
"loss": 1.0783, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 0.7558869701726845, |
|
"grad_norm": 0.3030723035335541, |
|
"learning_rate": 2.4489795918367347e-06, |
|
"loss": 1.0155, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.7566718995290423, |
|
"grad_norm": 0.28593307733535767, |
|
"learning_rate": 2.4411302982731556e-06, |
|
"loss": 1.0266, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 0.7574568288854003, |
|
"grad_norm": 0.2924596667289734, |
|
"learning_rate": 2.4332810047095766e-06, |
|
"loss": 1.0367, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.7582417582417582, |
|
"grad_norm": 0.30590420961380005, |
|
"learning_rate": 2.425431711145997e-06, |
|
"loss": 1.0675, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 0.7590266875981162, |
|
"grad_norm": 0.30233892798423767, |
|
"learning_rate": 2.4175824175824177e-06, |
|
"loss": 1.0837, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 0.7598116169544741, |
|
"grad_norm": 0.32924067974090576, |
|
"learning_rate": 2.4097331240188383e-06, |
|
"loss": 1.109, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 0.7605965463108321, |
|
"grad_norm": 0.29074007272720337, |
|
"learning_rate": 2.4018838304552593e-06, |
|
"loss": 1.0616, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 0.7613814756671899, |
|
"grad_norm": 0.29699182510375977, |
|
"learning_rate": 2.3940345368916803e-06, |
|
"loss": 1.0717, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.7621664050235479, |
|
"grad_norm": 0.3066222071647644, |
|
"learning_rate": 2.386185243328101e-06, |
|
"loss": 1.0731, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 0.7629513343799058, |
|
"grad_norm": 0.37514591217041016, |
|
"learning_rate": 2.3783359497645214e-06, |
|
"loss": 1.0188, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 0.7637362637362637, |
|
"grad_norm": 0.33850035071372986, |
|
"learning_rate": 2.370486656200942e-06, |
|
"loss": 1.1529, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 0.7645211930926217, |
|
"grad_norm": 0.2899448275566101, |
|
"learning_rate": 2.362637362637363e-06, |
|
"loss": 1.0658, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 0.7653061224489796, |
|
"grad_norm": 0.3076562285423279, |
|
"learning_rate": 2.3547880690737835e-06, |
|
"loss": 1.028, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.7660910518053375, |
|
"grad_norm": 0.3137950599193573, |
|
"learning_rate": 2.3469387755102044e-06, |
|
"loss": 1.0447, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 0.7668759811616954, |
|
"grad_norm": 0.30430495738983154, |
|
"learning_rate": 2.339089481946625e-06, |
|
"loss": 1.105, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 0.7676609105180534, |
|
"grad_norm": 0.38252878189086914, |
|
"learning_rate": 2.3312401883830456e-06, |
|
"loss": 1.1403, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 0.7684458398744113, |
|
"grad_norm": 0.30081915855407715, |
|
"learning_rate": 2.3233908948194665e-06, |
|
"loss": 1.109, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 0.7692307692307693, |
|
"grad_norm": 0.29588279128074646, |
|
"learning_rate": 2.315541601255887e-06, |
|
"loss": 1.0319, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.7700156985871272, |
|
"grad_norm": 0.3080218434333801, |
|
"learning_rate": 2.307692307692308e-06, |
|
"loss": 1.039, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 0.7708006279434851, |
|
"grad_norm": 0.2921229302883148, |
|
"learning_rate": 2.2998430141287286e-06, |
|
"loss": 1.0493, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 0.771585557299843, |
|
"grad_norm": 0.2984630763530731, |
|
"learning_rate": 2.291993720565149e-06, |
|
"loss": 1.0191, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 0.7723704866562009, |
|
"grad_norm": 0.28374841809272766, |
|
"learning_rate": 2.28414442700157e-06, |
|
"loss": 1.0103, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 0.7731554160125589, |
|
"grad_norm": 0.3007064163684845, |
|
"learning_rate": 2.2762951334379907e-06, |
|
"loss": 1.0387, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.7739403453689168, |
|
"grad_norm": 0.2927864193916321, |
|
"learning_rate": 2.2684458398744113e-06, |
|
"loss": 1.005, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 0.7747252747252747, |
|
"grad_norm": 0.3065125048160553, |
|
"learning_rate": 2.2605965463108323e-06, |
|
"loss": 1.0851, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 0.7755102040816326, |
|
"grad_norm": 0.27849113941192627, |
|
"learning_rate": 2.252747252747253e-06, |
|
"loss": 1.0251, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 0.7762951334379906, |
|
"grad_norm": 0.2948971688747406, |
|
"learning_rate": 2.244897959183674e-06, |
|
"loss": 1.0274, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 0.7770800627943485, |
|
"grad_norm": 0.3202616274356842, |
|
"learning_rate": 2.2370486656200944e-06, |
|
"loss": 1.0485, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.7778649921507065, |
|
"grad_norm": 0.3328281342983246, |
|
"learning_rate": 2.229199372056515e-06, |
|
"loss": 1.069, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 0.7786499215070644, |
|
"grad_norm": 0.3161095976829529, |
|
"learning_rate": 2.221350078492936e-06, |
|
"loss": 1.0848, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 0.7794348508634223, |
|
"grad_norm": 0.28824999928474426, |
|
"learning_rate": 2.2135007849293564e-06, |
|
"loss": 1.0654, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 0.7802197802197802, |
|
"grad_norm": 0.3667064309120178, |
|
"learning_rate": 2.2056514913657774e-06, |
|
"loss": 1.0116, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 0.7810047095761381, |
|
"grad_norm": 0.2977278530597687, |
|
"learning_rate": 2.197802197802198e-06, |
|
"loss": 1.0814, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.7817896389324961, |
|
"grad_norm": 0.29998522996902466, |
|
"learning_rate": 2.1899529042386185e-06, |
|
"loss": 1.0431, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 0.782574568288854, |
|
"grad_norm": 0.31411993503570557, |
|
"learning_rate": 2.1821036106750395e-06, |
|
"loss": 1.0926, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 0.783359497645212, |
|
"grad_norm": 0.29877665638923645, |
|
"learning_rate": 2.17425431711146e-06, |
|
"loss": 1.0989, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 0.7841444270015698, |
|
"grad_norm": 0.2992810010910034, |
|
"learning_rate": 2.166405023547881e-06, |
|
"loss": 1.0119, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 0.7849293563579278, |
|
"grad_norm": 0.2953478991985321, |
|
"learning_rate": 2.1585557299843016e-06, |
|
"loss": 1.0669, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.7857142857142857, |
|
"grad_norm": 0.29331153631210327, |
|
"learning_rate": 2.150706436420722e-06, |
|
"loss": 1.0554, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 0.7864992150706437, |
|
"grad_norm": 0.2879624664783478, |
|
"learning_rate": 2.1428571428571427e-06, |
|
"loss": 1.0212, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 0.7872841444270016, |
|
"grad_norm": 0.2884847819805145, |
|
"learning_rate": 2.1350078492935637e-06, |
|
"loss": 1.0376, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 0.7880690737833596, |
|
"grad_norm": 0.29227468371391296, |
|
"learning_rate": 2.1271585557299847e-06, |
|
"loss": 1.0444, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 0.7888540031397174, |
|
"grad_norm": 0.3019685745239258, |
|
"learning_rate": 2.1193092621664052e-06, |
|
"loss": 1.1065, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.7896389324960753, |
|
"grad_norm": 0.287661075592041, |
|
"learning_rate": 2.111459968602826e-06, |
|
"loss": 1.0141, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 0.7904238618524333, |
|
"grad_norm": 0.29692256450653076, |
|
"learning_rate": 2.1036106750392464e-06, |
|
"loss": 1.0183, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 0.7912087912087912, |
|
"grad_norm": 0.29249200224876404, |
|
"learning_rate": 2.0957613814756673e-06, |
|
"loss": 1.0084, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 0.7919937205651492, |
|
"grad_norm": 0.31126755475997925, |
|
"learning_rate": 2.0879120879120883e-06, |
|
"loss": 1.023, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 0.792778649921507, |
|
"grad_norm": 0.29185745120048523, |
|
"learning_rate": 2.080062794348509e-06, |
|
"loss": 0.9944, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.793563579277865, |
|
"grad_norm": 0.30141139030456543, |
|
"learning_rate": 2.0722135007849294e-06, |
|
"loss": 1.0885, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 0.7943485086342229, |
|
"grad_norm": 0.29048752784729004, |
|
"learning_rate": 2.06436420722135e-06, |
|
"loss": 1.0231, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 0.7951334379905809, |
|
"grad_norm": 0.3008350431919098, |
|
"learning_rate": 2.056514913657771e-06, |
|
"loss": 1.0388, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 0.7959183673469388, |
|
"grad_norm": 0.30450665950775146, |
|
"learning_rate": 2.048665620094192e-06, |
|
"loss": 1.0316, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 0.7967032967032966, |
|
"grad_norm": 0.34311988949775696, |
|
"learning_rate": 2.0408163265306125e-06, |
|
"loss": 1.1252, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.7974882260596546, |
|
"grad_norm": 0.28808602690696716, |
|
"learning_rate": 2.032967032967033e-06, |
|
"loss": 1.0348, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 0.7982731554160125, |
|
"grad_norm": 0.28176361322402954, |
|
"learning_rate": 2.0251177394034536e-06, |
|
"loss": 1.0127, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 0.7990580847723705, |
|
"grad_norm": 0.30243223905563354, |
|
"learning_rate": 2.0172684458398746e-06, |
|
"loss": 1.0655, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 0.7998430141287284, |
|
"grad_norm": 0.2991596460342407, |
|
"learning_rate": 2.0094191522762956e-06, |
|
"loss": 1.071, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 0.8006279434850864, |
|
"grad_norm": 0.31719931960105896, |
|
"learning_rate": 2.001569858712716e-06, |
|
"loss": 1.0703, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.8014128728414442, |
|
"grad_norm": 0.28864073753356934, |
|
"learning_rate": 1.9937205651491367e-06, |
|
"loss": 1.0865, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 0.8021978021978022, |
|
"grad_norm": 0.2995680272579193, |
|
"learning_rate": 1.9858712715855573e-06, |
|
"loss": 1.043, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 0.8029827315541601, |
|
"grad_norm": 0.30036595463752747, |
|
"learning_rate": 1.9780219780219782e-06, |
|
"loss": 1.0925, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 0.8037676609105181, |
|
"grad_norm": 0.2845197916030884, |
|
"learning_rate": 1.970172684458399e-06, |
|
"loss": 1.0439, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 0.804552590266876, |
|
"grad_norm": 0.29325416684150696, |
|
"learning_rate": 1.9623233908948198e-06, |
|
"loss": 1.047, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.8053375196232339, |
|
"grad_norm": 0.2978193163871765, |
|
"learning_rate": 1.9544740973312403e-06, |
|
"loss": 1.0581, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 0.8061224489795918, |
|
"grad_norm": 0.31198781728744507, |
|
"learning_rate": 1.946624803767661e-06, |
|
"loss": 1.0275, |
|
"step": 1027 |
|
}, |
|
{ |
|
"epoch": 0.8069073783359497, |
|
"grad_norm": 0.28849077224731445, |
|
"learning_rate": 1.938775510204082e-06, |
|
"loss": 1.0534, |
|
"step": 1028 |
|
}, |
|
{ |
|
"epoch": 0.8076923076923077, |
|
"grad_norm": 0.3035949766635895, |
|
"learning_rate": 1.9309262166405024e-06, |
|
"loss": 0.9924, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 0.8084772370486656, |
|
"grad_norm": 0.330161988735199, |
|
"learning_rate": 1.9230769230769234e-06, |
|
"loss": 1.0625, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.8092621664050236, |
|
"grad_norm": 0.28037184476852417, |
|
"learning_rate": 1.915227629513344e-06, |
|
"loss": 1.0381, |
|
"step": 1031 |
|
}, |
|
{ |
|
"epoch": 0.8100470957613815, |
|
"grad_norm": 0.3421080410480499, |
|
"learning_rate": 1.9073783359497645e-06, |
|
"loss": 1.0885, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 0.8108320251177394, |
|
"grad_norm": 0.3068152368068695, |
|
"learning_rate": 1.8995290423861853e-06, |
|
"loss": 1.0788, |
|
"step": 1033 |
|
}, |
|
{ |
|
"epoch": 0.8116169544740973, |
|
"grad_norm": 0.2811432182788849, |
|
"learning_rate": 1.891679748822606e-06, |
|
"loss": 1.0359, |
|
"step": 1034 |
|
}, |
|
{ |
|
"epoch": 0.8124018838304553, |
|
"grad_norm": 0.5453617572784424, |
|
"learning_rate": 1.883830455259027e-06, |
|
"loss": 1.0151, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.8131868131868132, |
|
"grad_norm": 0.28949666023254395, |
|
"learning_rate": 1.8759811616954476e-06, |
|
"loss": 1.0557, |
|
"step": 1036 |
|
}, |
|
{ |
|
"epoch": 0.8139717425431711, |
|
"grad_norm": 0.2827453017234802, |
|
"learning_rate": 1.8681318681318684e-06, |
|
"loss": 1.006, |
|
"step": 1037 |
|
}, |
|
{ |
|
"epoch": 0.814756671899529, |
|
"grad_norm": 0.2997809648513794, |
|
"learning_rate": 1.860282574568289e-06, |
|
"loss": 1.025, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 0.8155416012558869, |
|
"grad_norm": 0.3027696907520294, |
|
"learning_rate": 1.8524332810047097e-06, |
|
"loss": 1.0502, |
|
"step": 1039 |
|
}, |
|
{ |
|
"epoch": 0.8163265306122449, |
|
"grad_norm": 0.3114776909351349, |
|
"learning_rate": 1.8445839874411302e-06, |
|
"loss": 1.1017, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.8171114599686028, |
|
"grad_norm": 0.2964245676994324, |
|
"learning_rate": 1.8367346938775512e-06, |
|
"loss": 1.0149, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 0.8178963893249608, |
|
"grad_norm": 0.2923440933227539, |
|
"learning_rate": 1.828885400313972e-06, |
|
"loss": 1.073, |
|
"step": 1042 |
|
}, |
|
{ |
|
"epoch": 0.8186813186813187, |
|
"grad_norm": 0.2958196699619293, |
|
"learning_rate": 1.8210361067503925e-06, |
|
"loss": 1.0788, |
|
"step": 1043 |
|
}, |
|
{ |
|
"epoch": 0.8194662480376766, |
|
"grad_norm": 0.32801884412765503, |
|
"learning_rate": 1.8131868131868133e-06, |
|
"loss": 1.0645, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 0.8202511773940345, |
|
"grad_norm": 0.30840179324150085, |
|
"learning_rate": 1.8053375196232339e-06, |
|
"loss": 1.0397, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.8210361067503925, |
|
"grad_norm": 0.2844547927379608, |
|
"learning_rate": 1.7974882260596548e-06, |
|
"loss": 1.0534, |
|
"step": 1046 |
|
}, |
|
{ |
|
"epoch": 0.8218210361067504, |
|
"grad_norm": 0.31664690375328064, |
|
"learning_rate": 1.7896389324960756e-06, |
|
"loss": 1.0628, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 0.8226059654631083, |
|
"grad_norm": 0.3169183135032654, |
|
"learning_rate": 1.7817896389324962e-06, |
|
"loss": 1.0954, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 0.8233908948194663, |
|
"grad_norm": 0.2980157136917114, |
|
"learning_rate": 1.773940345368917e-06, |
|
"loss": 1.0459, |
|
"step": 1049 |
|
}, |
|
{ |
|
"epoch": 0.8241758241758241, |
|
"grad_norm": 0.3251033425331116, |
|
"learning_rate": 1.7660910518053375e-06, |
|
"loss": 1.1277, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.8249607535321821, |
|
"grad_norm": 0.3023360073566437, |
|
"learning_rate": 1.7582417582417585e-06, |
|
"loss": 1.0413, |
|
"step": 1051 |
|
}, |
|
{ |
|
"epoch": 0.82574568288854, |
|
"grad_norm": 0.33668410778045654, |
|
"learning_rate": 1.7503924646781792e-06, |
|
"loss": 1.0721, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 0.826530612244898, |
|
"grad_norm": 0.30133289098739624, |
|
"learning_rate": 1.7425431711145998e-06, |
|
"loss": 1.0245, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 0.8273155416012559, |
|
"grad_norm": 0.30766019225120544, |
|
"learning_rate": 1.7346938775510206e-06, |
|
"loss": 1.0795, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 0.8281004709576139, |
|
"grad_norm": 0.2983943521976471, |
|
"learning_rate": 1.7268445839874411e-06, |
|
"loss": 1.0699, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.8288854003139717, |
|
"grad_norm": 0.3063719570636749, |
|
"learning_rate": 1.7189952904238619e-06, |
|
"loss": 1.0624, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 0.8296703296703297, |
|
"grad_norm": 0.3908691704273224, |
|
"learning_rate": 1.7111459968602829e-06, |
|
"loss": 1.0461, |
|
"step": 1057 |
|
}, |
|
{ |
|
"epoch": 0.8304552590266876, |
|
"grad_norm": 0.30761247873306274, |
|
"learning_rate": 1.7032967032967034e-06, |
|
"loss": 1.0858, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 0.8312401883830455, |
|
"grad_norm": 0.2950478792190552, |
|
"learning_rate": 1.6954474097331242e-06, |
|
"loss": 1.0143, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 0.8320251177394035, |
|
"grad_norm": 0.30142104625701904, |
|
"learning_rate": 1.6875981161695448e-06, |
|
"loss": 1.0693, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.8328100470957613, |
|
"grad_norm": 0.30439862608909607, |
|
"learning_rate": 1.6797488226059655e-06, |
|
"loss": 1.0668, |
|
"step": 1061 |
|
}, |
|
{ |
|
"epoch": 0.8335949764521193, |
|
"grad_norm": 0.2978014349937439, |
|
"learning_rate": 1.6718995290423865e-06, |
|
"loss": 1.0782, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 0.8343799058084772, |
|
"grad_norm": 0.28805792331695557, |
|
"learning_rate": 1.664050235478807e-06, |
|
"loss": 1.0318, |
|
"step": 1063 |
|
}, |
|
{ |
|
"epoch": 0.8351648351648352, |
|
"grad_norm": 0.29518917202949524, |
|
"learning_rate": 1.6562009419152278e-06, |
|
"loss": 1.0475, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 0.8359497645211931, |
|
"grad_norm": 0.2990979552268982, |
|
"learning_rate": 1.6483516483516484e-06, |
|
"loss": 1.0352, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.8367346938775511, |
|
"grad_norm": 0.3014602065086365, |
|
"learning_rate": 1.6405023547880692e-06, |
|
"loss": 1.0568, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 0.8375196232339089, |
|
"grad_norm": 0.2860758602619171, |
|
"learning_rate": 1.6326530612244897e-06, |
|
"loss": 1.0387, |
|
"step": 1067 |
|
}, |
|
{ |
|
"epoch": 0.8383045525902669, |
|
"grad_norm": 0.293454647064209, |
|
"learning_rate": 1.6248037676609107e-06, |
|
"loss": 1.0444, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 0.8390894819466248, |
|
"grad_norm": 0.2916511297225952, |
|
"learning_rate": 1.6169544740973315e-06, |
|
"loss": 1.0512, |
|
"step": 1069 |
|
}, |
|
{ |
|
"epoch": 0.8398744113029827, |
|
"grad_norm": 0.30183103680610657, |
|
"learning_rate": 1.609105180533752e-06, |
|
"loss": 1.0257, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.8406593406593407, |
|
"grad_norm": 0.3018069267272949, |
|
"learning_rate": 1.6012558869701728e-06, |
|
"loss": 1.0665, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 0.8414442700156985, |
|
"grad_norm": 0.27915433049201965, |
|
"learning_rate": 1.5934065934065933e-06, |
|
"loss": 1.0397, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 0.8422291993720565, |
|
"grad_norm": 0.3076684772968292, |
|
"learning_rate": 1.5855572998430143e-06, |
|
"loss": 1.0981, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 0.8430141287284144, |
|
"grad_norm": 0.29367414116859436, |
|
"learning_rate": 1.577708006279435e-06, |
|
"loss": 1.0547, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 0.8437990580847724, |
|
"grad_norm": 0.2984970211982727, |
|
"learning_rate": 1.5698587127158556e-06, |
|
"loss": 1.0494, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.8445839874411303, |
|
"grad_norm": 0.29925835132598877, |
|
"learning_rate": 1.5620094191522764e-06, |
|
"loss": 1.0686, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 0.8453689167974883, |
|
"grad_norm": 0.2967824935913086, |
|
"learning_rate": 1.554160125588697e-06, |
|
"loss": 1.1013, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 0.8461538461538461, |
|
"grad_norm": 0.30394864082336426, |
|
"learning_rate": 1.546310832025118e-06, |
|
"loss": 1.0908, |
|
"step": 1078 |
|
}, |
|
{ |
|
"epoch": 0.8469387755102041, |
|
"grad_norm": 0.3596284091472626, |
|
"learning_rate": 1.5384615384615387e-06, |
|
"loss": 1.0496, |
|
"step": 1079 |
|
}, |
|
{ |
|
"epoch": 0.847723704866562, |
|
"grad_norm": 0.30267584323883057, |
|
"learning_rate": 1.5306122448979593e-06, |
|
"loss": 1.063, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.8485086342229199, |
|
"grad_norm": 0.2946220934391022, |
|
"learning_rate": 1.52276295133438e-06, |
|
"loss": 1.056, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 0.8492935635792779, |
|
"grad_norm": 0.2994774281978607, |
|
"learning_rate": 1.5149136577708006e-06, |
|
"loss": 1.055, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 0.8500784929356358, |
|
"grad_norm": 0.2964215576648712, |
|
"learning_rate": 1.5070643642072214e-06, |
|
"loss": 1.0636, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 0.8508634222919937, |
|
"grad_norm": 0.313342422246933, |
|
"learning_rate": 1.4992150706436423e-06, |
|
"loss": 1.0341, |
|
"step": 1084 |
|
}, |
|
{ |
|
"epoch": 0.8516483516483516, |
|
"grad_norm": 0.30197572708129883, |
|
"learning_rate": 1.491365777080063e-06, |
|
"loss": 1.0172, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.8524332810047096, |
|
"grad_norm": 0.3112085461616516, |
|
"learning_rate": 1.4835164835164837e-06, |
|
"loss": 1.0727, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 0.8532182103610675, |
|
"grad_norm": 0.3003901243209839, |
|
"learning_rate": 1.4756671899529042e-06, |
|
"loss": 1.0527, |
|
"step": 1087 |
|
}, |
|
{ |
|
"epoch": 0.8540031397174255, |
|
"grad_norm": 0.2971203327178955, |
|
"learning_rate": 1.467817896389325e-06, |
|
"loss": 1.041, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 0.8547880690737834, |
|
"grad_norm": 0.3106226921081543, |
|
"learning_rate": 1.459968602825746e-06, |
|
"loss": 1.1165, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 0.8555729984301413, |
|
"grad_norm": 0.31148043274879456, |
|
"learning_rate": 1.4521193092621665e-06, |
|
"loss": 1.0847, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.8563579277864992, |
|
"grad_norm": 0.29365935921669006, |
|
"learning_rate": 1.4442700156985873e-06, |
|
"loss": 1.013, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 0.8571428571428571, |
|
"grad_norm": 0.30795639753341675, |
|
"learning_rate": 1.4364207221350079e-06, |
|
"loss": 1.0386, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 0.8579277864992151, |
|
"grad_norm": 0.2929840087890625, |
|
"learning_rate": 1.4285714285714286e-06, |
|
"loss": 1.036, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 0.858712715855573, |
|
"grad_norm": 0.31012848019599915, |
|
"learning_rate": 1.4207221350078492e-06, |
|
"loss": 1.0885, |
|
"step": 1094 |
|
}, |
|
{ |
|
"epoch": 0.859497645211931, |
|
"grad_norm": 0.28400346636772156, |
|
"learning_rate": 1.4128728414442702e-06, |
|
"loss": 1.0615, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.8602825745682888, |
|
"grad_norm": 0.28853991627693176, |
|
"learning_rate": 1.405023547880691e-06, |
|
"loss": 1.0612, |
|
"step": 1096 |
|
}, |
|
{ |
|
"epoch": 0.8610675039246468, |
|
"grad_norm": 0.28974905610084534, |
|
"learning_rate": 1.3971742543171115e-06, |
|
"loss": 1.0506, |
|
"step": 1097 |
|
}, |
|
{ |
|
"epoch": 0.8618524332810047, |
|
"grad_norm": 0.3157826066017151, |
|
"learning_rate": 1.3893249607535323e-06, |
|
"loss": 1.0691, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 0.8626373626373627, |
|
"grad_norm": 0.28993314504623413, |
|
"learning_rate": 1.3814756671899528e-06, |
|
"loss": 1.0108, |
|
"step": 1099 |
|
}, |
|
{ |
|
"epoch": 0.8634222919937206, |
|
"grad_norm": 0.282035231590271, |
|
"learning_rate": 1.3736263736263738e-06, |
|
"loss": 1.0208, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.8642072213500785, |
|
"grad_norm": 0.30081960558891296, |
|
"learning_rate": 1.3657770800627946e-06, |
|
"loss": 1.0423, |
|
"step": 1101 |
|
}, |
|
{ |
|
"epoch": 0.8649921507064364, |
|
"grad_norm": 0.3326198160648346, |
|
"learning_rate": 1.3579277864992151e-06, |
|
"loss": 1.0326, |
|
"step": 1102 |
|
}, |
|
{ |
|
"epoch": 0.8657770800627943, |
|
"grad_norm": 0.2900926470756531, |
|
"learning_rate": 1.3500784929356359e-06, |
|
"loss": 1.0525, |
|
"step": 1103 |
|
}, |
|
{ |
|
"epoch": 0.8665620094191523, |
|
"grad_norm": 0.28752028942108154, |
|
"learning_rate": 1.3422291993720564e-06, |
|
"loss": 1.0074, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 0.8673469387755102, |
|
"grad_norm": 0.2825300395488739, |
|
"learning_rate": 1.3343799058084774e-06, |
|
"loss": 1.0223, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.8681318681318682, |
|
"grad_norm": 0.30561885237693787, |
|
"learning_rate": 1.3265306122448982e-06, |
|
"loss": 1.0464, |
|
"step": 1106 |
|
}, |
|
{ |
|
"epoch": 0.868916797488226, |
|
"grad_norm": 0.28518933057785034, |
|
"learning_rate": 1.3186813186813187e-06, |
|
"loss": 1.0464, |
|
"step": 1107 |
|
}, |
|
{ |
|
"epoch": 0.869701726844584, |
|
"grad_norm": 0.2933896780014038, |
|
"learning_rate": 1.3108320251177395e-06, |
|
"loss": 1.0177, |
|
"step": 1108 |
|
}, |
|
{ |
|
"epoch": 0.8704866562009419, |
|
"grad_norm": 0.30531638860702515, |
|
"learning_rate": 1.30298273155416e-06, |
|
"loss": 0.9894, |
|
"step": 1109 |
|
}, |
|
{ |
|
"epoch": 0.8712715855572999, |
|
"grad_norm": 0.2906123995780945, |
|
"learning_rate": 1.2951334379905808e-06, |
|
"loss": 1.0403, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.8720565149136578, |
|
"grad_norm": 0.29348504543304443, |
|
"learning_rate": 1.2872841444270018e-06, |
|
"loss": 1.0439, |
|
"step": 1111 |
|
}, |
|
{ |
|
"epoch": 0.8728414442700158, |
|
"grad_norm": 0.3133821189403534, |
|
"learning_rate": 1.2794348508634224e-06, |
|
"loss": 1.0984, |
|
"step": 1112 |
|
}, |
|
{ |
|
"epoch": 0.8736263736263736, |
|
"grad_norm": 0.2935754358768463, |
|
"learning_rate": 1.2715855572998431e-06, |
|
"loss": 1.0538, |
|
"step": 1113 |
|
}, |
|
{ |
|
"epoch": 0.8744113029827315, |
|
"grad_norm": 0.484567791223526, |
|
"learning_rate": 1.2637362637362637e-06, |
|
"loss": 0.9939, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.8751962323390895, |
|
"grad_norm": 0.2972055673599243, |
|
"learning_rate": 1.2558869701726845e-06, |
|
"loss": 1.0524, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.8759811616954474, |
|
"grad_norm": 0.3031924366950989, |
|
"learning_rate": 1.2480376766091052e-06, |
|
"loss": 1.0597, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 0.8767660910518054, |
|
"grad_norm": 0.2977665662765503, |
|
"learning_rate": 1.240188383045526e-06, |
|
"loss": 1.0796, |
|
"step": 1117 |
|
}, |
|
{ |
|
"epoch": 0.8775510204081632, |
|
"grad_norm": 0.3066113591194153, |
|
"learning_rate": 1.2323390894819468e-06, |
|
"loss": 1.0741, |
|
"step": 1118 |
|
}, |
|
{ |
|
"epoch": 0.8783359497645212, |
|
"grad_norm": 0.30936139822006226, |
|
"learning_rate": 1.2244897959183673e-06, |
|
"loss": 1.0501, |
|
"step": 1119 |
|
}, |
|
{ |
|
"epoch": 0.8791208791208791, |
|
"grad_norm": 0.32315778732299805, |
|
"learning_rate": 1.2166405023547883e-06, |
|
"loss": 1.0674, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.8799058084772371, |
|
"grad_norm": 0.3213968276977539, |
|
"learning_rate": 1.2087912087912089e-06, |
|
"loss": 1.079, |
|
"step": 1121 |
|
}, |
|
{ |
|
"epoch": 0.880690737833595, |
|
"grad_norm": 0.30559977889060974, |
|
"learning_rate": 1.2009419152276296e-06, |
|
"loss": 1.0145, |
|
"step": 1122 |
|
}, |
|
{ |
|
"epoch": 0.8814756671899528, |
|
"grad_norm": 0.32773932814598083, |
|
"learning_rate": 1.1930926216640504e-06, |
|
"loss": 1.0227, |
|
"step": 1123 |
|
}, |
|
{ |
|
"epoch": 0.8822605965463108, |
|
"grad_norm": 0.2936771810054779, |
|
"learning_rate": 1.185243328100471e-06, |
|
"loss": 1.0222, |
|
"step": 1124 |
|
}, |
|
{ |
|
"epoch": 0.8830455259026687, |
|
"grad_norm": 0.3083963096141815, |
|
"learning_rate": 1.1773940345368917e-06, |
|
"loss": 1.0424, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.8838304552590267, |
|
"grad_norm": 0.3056409955024719, |
|
"learning_rate": 1.1695447409733125e-06, |
|
"loss": 1.087, |
|
"step": 1126 |
|
}, |
|
{ |
|
"epoch": 0.8846153846153846, |
|
"grad_norm": 0.2972738444805145, |
|
"learning_rate": 1.1616954474097333e-06, |
|
"loss": 1.034, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 0.8854003139717426, |
|
"grad_norm": 0.28523531556129456, |
|
"learning_rate": 1.153846153846154e-06, |
|
"loss": 1.0371, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 0.8861852433281004, |
|
"grad_norm": 0.3151058852672577, |
|
"learning_rate": 1.1459968602825746e-06, |
|
"loss": 1.0826, |
|
"step": 1129 |
|
}, |
|
{ |
|
"epoch": 0.8869701726844584, |
|
"grad_norm": 0.2846231460571289, |
|
"learning_rate": 1.1381475667189954e-06, |
|
"loss": 1.0419, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.8877551020408163, |
|
"grad_norm": 0.3189791738986969, |
|
"learning_rate": 1.1302982731554161e-06, |
|
"loss": 1.0161, |
|
"step": 1131 |
|
}, |
|
{ |
|
"epoch": 0.8885400313971743, |
|
"grad_norm": 0.29699793457984924, |
|
"learning_rate": 1.122448979591837e-06, |
|
"loss": 1.0944, |
|
"step": 1132 |
|
}, |
|
{ |
|
"epoch": 0.8893249607535322, |
|
"grad_norm": 0.31134846806526184, |
|
"learning_rate": 1.1145996860282575e-06, |
|
"loss": 1.0405, |
|
"step": 1133 |
|
}, |
|
{ |
|
"epoch": 0.8901098901098901, |
|
"grad_norm": 0.3218204975128174, |
|
"learning_rate": 1.1067503924646782e-06, |
|
"loss": 1.0721, |
|
"step": 1134 |
|
}, |
|
{ |
|
"epoch": 0.890894819466248, |
|
"grad_norm": 0.2882716655731201, |
|
"learning_rate": 1.098901098901099e-06, |
|
"loss": 1.0527, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.8916797488226059, |
|
"grad_norm": 0.2993222177028656, |
|
"learning_rate": 1.0910518053375198e-06, |
|
"loss": 1.0707, |
|
"step": 1136 |
|
}, |
|
{ |
|
"epoch": 0.8924646781789639, |
|
"grad_norm": 0.29466983675956726, |
|
"learning_rate": 1.0832025117739405e-06, |
|
"loss": 1.0645, |
|
"step": 1137 |
|
}, |
|
{ |
|
"epoch": 0.8932496075353218, |
|
"grad_norm": 0.29492950439453125, |
|
"learning_rate": 1.075353218210361e-06, |
|
"loss": 1.0756, |
|
"step": 1138 |
|
}, |
|
{ |
|
"epoch": 0.8940345368916798, |
|
"grad_norm": 0.2911151647567749, |
|
"learning_rate": 1.0675039246467819e-06, |
|
"loss": 1.0676, |
|
"step": 1139 |
|
}, |
|
{ |
|
"epoch": 0.8948194662480377, |
|
"grad_norm": 0.3016468286514282, |
|
"learning_rate": 1.0596546310832026e-06, |
|
"loss": 1.0752, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.8956043956043956, |
|
"grad_norm": 0.2953685522079468, |
|
"learning_rate": 1.0518053375196232e-06, |
|
"loss": 1.001, |
|
"step": 1141 |
|
}, |
|
{ |
|
"epoch": 0.8963893249607535, |
|
"grad_norm": 0.2959998548030853, |
|
"learning_rate": 1.0439560439560442e-06, |
|
"loss": 1.0695, |
|
"step": 1142 |
|
}, |
|
{ |
|
"epoch": 0.8971742543171115, |
|
"grad_norm": 0.29669925570487976, |
|
"learning_rate": 1.0361067503924647e-06, |
|
"loss": 1.0214, |
|
"step": 1143 |
|
}, |
|
{ |
|
"epoch": 0.8979591836734694, |
|
"grad_norm": 0.28717902302742004, |
|
"learning_rate": 1.0282574568288855e-06, |
|
"loss": 1.0687, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 0.8987441130298273, |
|
"grad_norm": 0.29382869601249695, |
|
"learning_rate": 1.0204081632653063e-06, |
|
"loss": 1.0213, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.8995290423861853, |
|
"grad_norm": 0.2865571081638336, |
|
"learning_rate": 1.0125588697017268e-06, |
|
"loss": 1.0308, |
|
"step": 1146 |
|
}, |
|
{ |
|
"epoch": 0.9003139717425431, |
|
"grad_norm": 0.3298538029193878, |
|
"learning_rate": 1.0047095761381478e-06, |
|
"loss": 1.0949, |
|
"step": 1147 |
|
}, |
|
{ |
|
"epoch": 0.9010989010989011, |
|
"grad_norm": 0.2812543511390686, |
|
"learning_rate": 9.968602825745683e-07, |
|
"loss": 1.0392, |
|
"step": 1148 |
|
}, |
|
{ |
|
"epoch": 0.901883830455259, |
|
"grad_norm": 0.29315754771232605, |
|
"learning_rate": 9.890109890109891e-07, |
|
"loss": 1.0201, |
|
"step": 1149 |
|
}, |
|
{ |
|
"epoch": 0.902668759811617, |
|
"grad_norm": 0.3410029411315918, |
|
"learning_rate": 9.811616954474099e-07, |
|
"loss": 1.1048, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.9034536891679749, |
|
"grad_norm": 0.2857743203639984, |
|
"learning_rate": 9.733124018838304e-07, |
|
"loss": 1.0611, |
|
"step": 1151 |
|
}, |
|
{ |
|
"epoch": 0.9042386185243328, |
|
"grad_norm": 0.29381293058395386, |
|
"learning_rate": 9.654631083202512e-07, |
|
"loss": 1.0146, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 0.9050235478806907, |
|
"grad_norm": 0.29993733763694763, |
|
"learning_rate": 9.57613814756672e-07, |
|
"loss": 1.0831, |
|
"step": 1153 |
|
}, |
|
{ |
|
"epoch": 0.9058084772370487, |
|
"grad_norm": 0.2960602343082428, |
|
"learning_rate": 9.497645211930926e-07, |
|
"loss": 1.0313, |
|
"step": 1154 |
|
}, |
|
{ |
|
"epoch": 0.9065934065934066, |
|
"grad_norm": 0.3170572817325592, |
|
"learning_rate": 9.419152276295135e-07, |
|
"loss": 1.0092, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.9073783359497645, |
|
"grad_norm": 0.4631412625312805, |
|
"learning_rate": 9.340659340659342e-07, |
|
"loss": 1.0447, |
|
"step": 1156 |
|
}, |
|
{ |
|
"epoch": 0.9081632653061225, |
|
"grad_norm": 0.2946299910545349, |
|
"learning_rate": 9.262166405023548e-07, |
|
"loss": 1.0705, |
|
"step": 1157 |
|
}, |
|
{ |
|
"epoch": 0.9089481946624803, |
|
"grad_norm": 0.30375024676322937, |
|
"learning_rate": 9.183673469387756e-07, |
|
"loss": 1.0704, |
|
"step": 1158 |
|
}, |
|
{ |
|
"epoch": 0.9097331240188383, |
|
"grad_norm": 0.2881094515323639, |
|
"learning_rate": 9.105180533751963e-07, |
|
"loss": 1.0145, |
|
"step": 1159 |
|
}, |
|
{ |
|
"epoch": 0.9105180533751962, |
|
"grad_norm": 0.29087066650390625, |
|
"learning_rate": 9.026687598116169e-07, |
|
"loss": 1.0452, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.9113029827315542, |
|
"grad_norm": 0.29212790727615356, |
|
"learning_rate": 8.948194662480378e-07, |
|
"loss": 1.0573, |
|
"step": 1161 |
|
}, |
|
{ |
|
"epoch": 0.9120879120879121, |
|
"grad_norm": 0.2992939352989197, |
|
"learning_rate": 8.869701726844585e-07, |
|
"loss": 1.0425, |
|
"step": 1162 |
|
}, |
|
{ |
|
"epoch": 0.9128728414442701, |
|
"grad_norm": 0.29093456268310547, |
|
"learning_rate": 8.791208791208792e-07, |
|
"loss": 1.0422, |
|
"step": 1163 |
|
}, |
|
{ |
|
"epoch": 0.9136577708006279, |
|
"grad_norm": 0.2929815351963043, |
|
"learning_rate": 8.712715855572999e-07, |
|
"loss": 1.0586, |
|
"step": 1164 |
|
}, |
|
{ |
|
"epoch": 0.9144427001569859, |
|
"grad_norm": 0.36663711071014404, |
|
"learning_rate": 8.634222919937206e-07, |
|
"loss": 1.0573, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.9152276295133438, |
|
"grad_norm": 0.3045317232608795, |
|
"learning_rate": 8.555729984301414e-07, |
|
"loss": 1.1, |
|
"step": 1166 |
|
}, |
|
{ |
|
"epoch": 0.9160125588697017, |
|
"grad_norm": 0.30183184146881104, |
|
"learning_rate": 8.477237048665621e-07, |
|
"loss": 1.036, |
|
"step": 1167 |
|
}, |
|
{ |
|
"epoch": 0.9167974882260597, |
|
"grad_norm": 0.3052210509777069, |
|
"learning_rate": 8.398744113029828e-07, |
|
"loss": 1.0897, |
|
"step": 1168 |
|
}, |
|
{ |
|
"epoch": 0.9175824175824175, |
|
"grad_norm": 0.3007582724094391, |
|
"learning_rate": 8.320251177394035e-07, |
|
"loss": 1.0259, |
|
"step": 1169 |
|
}, |
|
{ |
|
"epoch": 0.9183673469387755, |
|
"grad_norm": 0.3091794550418854, |
|
"learning_rate": 8.241758241758242e-07, |
|
"loss": 1.0354, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.9191522762951334, |
|
"grad_norm": 0.30170249938964844, |
|
"learning_rate": 8.163265306122449e-07, |
|
"loss": 1.025, |
|
"step": 1171 |
|
}, |
|
{ |
|
"epoch": 0.9199372056514914, |
|
"grad_norm": 0.5053988695144653, |
|
"learning_rate": 8.084772370486657e-07, |
|
"loss": 1.087, |
|
"step": 1172 |
|
}, |
|
{ |
|
"epoch": 0.9207221350078493, |
|
"grad_norm": 0.3013533055782318, |
|
"learning_rate": 8.006279434850864e-07, |
|
"loss": 1.0278, |
|
"step": 1173 |
|
}, |
|
{ |
|
"epoch": 0.9215070643642073, |
|
"grad_norm": 0.3028901517391205, |
|
"learning_rate": 7.927786499215072e-07, |
|
"loss": 1.0686, |
|
"step": 1174 |
|
}, |
|
{ |
|
"epoch": 0.9222919937205651, |
|
"grad_norm": 0.28716418147087097, |
|
"learning_rate": 7.849293563579278e-07, |
|
"loss": 1.0335, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.9230769230769231, |
|
"grad_norm": 0.3054925501346588, |
|
"learning_rate": 7.770800627943485e-07, |
|
"loss": 1.0697, |
|
"step": 1176 |
|
}, |
|
{ |
|
"epoch": 0.923861852433281, |
|
"grad_norm": 0.3063417673110962, |
|
"learning_rate": 7.692307692307694e-07, |
|
"loss": 1.061, |
|
"step": 1177 |
|
}, |
|
{ |
|
"epoch": 0.9246467817896389, |
|
"grad_norm": 0.28995341062545776, |
|
"learning_rate": 7.6138147566719e-07, |
|
"loss": 1.0103, |
|
"step": 1178 |
|
}, |
|
{ |
|
"epoch": 0.9254317111459969, |
|
"grad_norm": 0.2932472825050354, |
|
"learning_rate": 7.535321821036107e-07, |
|
"loss": 1.0164, |
|
"step": 1179 |
|
}, |
|
{ |
|
"epoch": 0.9262166405023547, |
|
"grad_norm": 0.3210296034812927, |
|
"learning_rate": 7.456828885400315e-07, |
|
"loss": 1.118, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.9270015698587127, |
|
"grad_norm": 0.30883345007896423, |
|
"learning_rate": 7.378335949764521e-07, |
|
"loss": 1.0462, |
|
"step": 1181 |
|
}, |
|
{ |
|
"epoch": 0.9277864992150706, |
|
"grad_norm": 0.2974553406238556, |
|
"learning_rate": 7.29984301412873e-07, |
|
"loss": 1.0685, |
|
"step": 1182 |
|
}, |
|
{ |
|
"epoch": 0.9285714285714286, |
|
"grad_norm": 0.2933749258518219, |
|
"learning_rate": 7.221350078492937e-07, |
|
"loss": 1.0296, |
|
"step": 1183 |
|
}, |
|
{ |
|
"epoch": 0.9293563579277865, |
|
"grad_norm": 0.2952398657798767, |
|
"learning_rate": 7.142857142857143e-07, |
|
"loss": 1.015, |
|
"step": 1184 |
|
}, |
|
{ |
|
"epoch": 0.9301412872841445, |
|
"grad_norm": 0.3025822043418884, |
|
"learning_rate": 7.064364207221351e-07, |
|
"loss": 1.0847, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.9309262166405023, |
|
"grad_norm": 0.2824312448501587, |
|
"learning_rate": 6.985871271585557e-07, |
|
"loss": 0.9973, |
|
"step": 1186 |
|
}, |
|
{ |
|
"epoch": 0.9317111459968603, |
|
"grad_norm": 0.29934030771255493, |
|
"learning_rate": 6.907378335949764e-07, |
|
"loss": 1.0566, |
|
"step": 1187 |
|
}, |
|
{ |
|
"epoch": 0.9324960753532182, |
|
"grad_norm": 0.29235753417015076, |
|
"learning_rate": 6.828885400313973e-07, |
|
"loss": 0.9836, |
|
"step": 1188 |
|
}, |
|
{ |
|
"epoch": 0.9332810047095761, |
|
"grad_norm": 0.3037624955177307, |
|
"learning_rate": 6.750392464678179e-07, |
|
"loss": 1.0722, |
|
"step": 1189 |
|
}, |
|
{ |
|
"epoch": 0.9340659340659341, |
|
"grad_norm": 0.2828160524368286, |
|
"learning_rate": 6.671899529042387e-07, |
|
"loss": 1.0393, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.934850863422292, |
|
"grad_norm": 0.28773877024650574, |
|
"learning_rate": 6.593406593406594e-07, |
|
"loss": 1.0417, |
|
"step": 1191 |
|
}, |
|
{ |
|
"epoch": 0.9356357927786499, |
|
"grad_norm": 0.2887474596500397, |
|
"learning_rate": 6.5149136577708e-07, |
|
"loss": 1.0321, |
|
"step": 1192 |
|
}, |
|
{ |
|
"epoch": 0.9364207221350078, |
|
"grad_norm": 0.2963563799858093, |
|
"learning_rate": 6.436420722135009e-07, |
|
"loss": 1.0139, |
|
"step": 1193 |
|
}, |
|
{ |
|
"epoch": 0.9372056514913658, |
|
"grad_norm": 0.2964331805706024, |
|
"learning_rate": 6.357927786499216e-07, |
|
"loss": 1.0662, |
|
"step": 1194 |
|
}, |
|
{ |
|
"epoch": 0.9379905808477237, |
|
"grad_norm": 0.29119017720222473, |
|
"learning_rate": 6.279434850863422e-07, |
|
"loss": 1.0797, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.9387755102040817, |
|
"grad_norm": 0.32927101850509644, |
|
"learning_rate": 6.20094191522763e-07, |
|
"loss": 1.0963, |
|
"step": 1196 |
|
}, |
|
{ |
|
"epoch": 0.9395604395604396, |
|
"grad_norm": 0.2921772003173828, |
|
"learning_rate": 6.122448979591837e-07, |
|
"loss": 1.0088, |
|
"step": 1197 |
|
}, |
|
{ |
|
"epoch": 0.9403453689167975, |
|
"grad_norm": 0.28854402899742126, |
|
"learning_rate": 6.043956043956044e-07, |
|
"loss": 1.0501, |
|
"step": 1198 |
|
}, |
|
{ |
|
"epoch": 0.9411302982731554, |
|
"grad_norm": 0.30962881445884705, |
|
"learning_rate": 5.965463108320252e-07, |
|
"loss": 1.1162, |
|
"step": 1199 |
|
}, |
|
{ |
|
"epoch": 0.9419152276295133, |
|
"grad_norm": 0.2930225729942322, |
|
"learning_rate": 5.886970172684459e-07, |
|
"loss": 1.0008, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.9427001569858713, |
|
"grad_norm": 0.3310118317604065, |
|
"learning_rate": 5.808477237048666e-07, |
|
"loss": 1.0714, |
|
"step": 1201 |
|
}, |
|
{ |
|
"epoch": 0.9434850863422292, |
|
"grad_norm": 0.300510048866272, |
|
"learning_rate": 5.729984301412873e-07, |
|
"loss": 1.054, |
|
"step": 1202 |
|
}, |
|
{ |
|
"epoch": 0.9442700156985872, |
|
"grad_norm": 0.3338243067264557, |
|
"learning_rate": 5.651491365777081e-07, |
|
"loss": 1.0849, |
|
"step": 1203 |
|
}, |
|
{ |
|
"epoch": 0.945054945054945, |
|
"grad_norm": 0.29972943663597107, |
|
"learning_rate": 5.572998430141287e-07, |
|
"loss": 1.0148, |
|
"step": 1204 |
|
}, |
|
{ |
|
"epoch": 0.945839874411303, |
|
"grad_norm": 0.29417991638183594, |
|
"learning_rate": 5.494505494505495e-07, |
|
"loss": 1.0307, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.9466248037676609, |
|
"grad_norm": 0.2904272675514221, |
|
"learning_rate": 5.416012558869703e-07, |
|
"loss": 1.0504, |
|
"step": 1206 |
|
}, |
|
{ |
|
"epoch": 0.9474097331240189, |
|
"grad_norm": 0.37245574593544006, |
|
"learning_rate": 5.337519623233909e-07, |
|
"loss": 1.077, |
|
"step": 1207 |
|
}, |
|
{ |
|
"epoch": 0.9481946624803768, |
|
"grad_norm": 0.3075472414493561, |
|
"learning_rate": 5.259026687598116e-07, |
|
"loss": 1.0483, |
|
"step": 1208 |
|
}, |
|
{ |
|
"epoch": 0.9489795918367347, |
|
"grad_norm": 0.29694482684135437, |
|
"learning_rate": 5.180533751962324e-07, |
|
"loss": 1.0239, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 0.9497645211930926, |
|
"grad_norm": 0.3054351806640625, |
|
"learning_rate": 5.102040816326531e-07, |
|
"loss": 1.0446, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.9505494505494505, |
|
"grad_norm": 0.30204030871391296, |
|
"learning_rate": 5.023547880690739e-07, |
|
"loss": 1.0195, |
|
"step": 1211 |
|
}, |
|
{ |
|
"epoch": 0.9513343799058085, |
|
"grad_norm": 0.29818445444107056, |
|
"learning_rate": 4.945054945054946e-07, |
|
"loss": 1.0322, |
|
"step": 1212 |
|
}, |
|
{ |
|
"epoch": 0.9521193092621664, |
|
"grad_norm": 0.4030686914920807, |
|
"learning_rate": 4.866562009419152e-07, |
|
"loss": 0.9908, |
|
"step": 1213 |
|
}, |
|
{ |
|
"epoch": 0.9529042386185244, |
|
"grad_norm": 0.2839055061340332, |
|
"learning_rate": 4.78806907378336e-07, |
|
"loss": 1.0366, |
|
"step": 1214 |
|
}, |
|
{ |
|
"epoch": 0.9536891679748822, |
|
"grad_norm": 0.29423198103904724, |
|
"learning_rate": 4.7095761381475676e-07, |
|
"loss": 1.038, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.9544740973312402, |
|
"grad_norm": 0.30914145708084106, |
|
"learning_rate": 4.631083202511774e-07, |
|
"loss": 1.1039, |
|
"step": 1216 |
|
}, |
|
{ |
|
"epoch": 0.9552590266875981, |
|
"grad_norm": 0.3080761730670929, |
|
"learning_rate": 4.5525902668759813e-07, |
|
"loss": 1.1072, |
|
"step": 1217 |
|
}, |
|
{ |
|
"epoch": 0.9560439560439561, |
|
"grad_norm": 0.3054615259170532, |
|
"learning_rate": 4.474097331240189e-07, |
|
"loss": 1.0865, |
|
"step": 1218 |
|
}, |
|
{ |
|
"epoch": 0.956828885400314, |
|
"grad_norm": 0.3009425103664398, |
|
"learning_rate": 4.395604395604396e-07, |
|
"loss": 1.0368, |
|
"step": 1219 |
|
}, |
|
{ |
|
"epoch": 0.957613814756672, |
|
"grad_norm": 0.28634223341941833, |
|
"learning_rate": 4.317111459968603e-07, |
|
"loss": 1.0526, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.9583987441130298, |
|
"grad_norm": 0.2902422547340393, |
|
"learning_rate": 4.2386185243328105e-07, |
|
"loss": 1.0402, |
|
"step": 1221 |
|
}, |
|
{ |
|
"epoch": 0.9591836734693877, |
|
"grad_norm": 0.31759124994277954, |
|
"learning_rate": 4.1601255886970176e-07, |
|
"loss": 1.0519, |
|
"step": 1222 |
|
}, |
|
{ |
|
"epoch": 0.9599686028257457, |
|
"grad_norm": 0.4426363408565521, |
|
"learning_rate": 4.0816326530612243e-07, |
|
"loss": 1.017, |
|
"step": 1223 |
|
}, |
|
{ |
|
"epoch": 0.9607535321821036, |
|
"grad_norm": 0.3140206038951874, |
|
"learning_rate": 4.003139717425432e-07, |
|
"loss": 1.0142, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 0.9615384615384616, |
|
"grad_norm": 0.29569703340530396, |
|
"learning_rate": 3.924646781789639e-07, |
|
"loss": 1.0528, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.9623233908948194, |
|
"grad_norm": 0.30597856640815735, |
|
"learning_rate": 3.846153846153847e-07, |
|
"loss": 1.0253, |
|
"step": 1226 |
|
}, |
|
{ |
|
"epoch": 0.9631083202511774, |
|
"grad_norm": 0.3235307037830353, |
|
"learning_rate": 3.7676609105180534e-07, |
|
"loss": 1.1179, |
|
"step": 1227 |
|
}, |
|
{ |
|
"epoch": 0.9638932496075353, |
|
"grad_norm": 0.2932690680027008, |
|
"learning_rate": 3.6891679748822606e-07, |
|
"loss": 1.0423, |
|
"step": 1228 |
|
}, |
|
{ |
|
"epoch": 0.9646781789638933, |
|
"grad_norm": 0.3032452166080475, |
|
"learning_rate": 3.610675039246468e-07, |
|
"loss": 1.0115, |
|
"step": 1229 |
|
}, |
|
{ |
|
"epoch": 0.9654631083202512, |
|
"grad_norm": 0.3229339122772217, |
|
"learning_rate": 3.5321821036106754e-07, |
|
"loss": 1.0766, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.966248037676609, |
|
"grad_norm": 0.3844963312149048, |
|
"learning_rate": 3.453689167974882e-07, |
|
"loss": 1.1032, |
|
"step": 1231 |
|
}, |
|
{ |
|
"epoch": 0.967032967032967, |
|
"grad_norm": 0.2927655279636383, |
|
"learning_rate": 3.3751962323390897e-07, |
|
"loss": 1.052, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 0.9678178963893249, |
|
"grad_norm": 0.3003545105457306, |
|
"learning_rate": 3.296703296703297e-07, |
|
"loss": 1.0825, |
|
"step": 1233 |
|
}, |
|
{ |
|
"epoch": 0.9686028257456829, |
|
"grad_norm": 0.32615581154823303, |
|
"learning_rate": 3.2182103610675046e-07, |
|
"loss": 1.0915, |
|
"step": 1234 |
|
}, |
|
{ |
|
"epoch": 0.9693877551020408, |
|
"grad_norm": 0.31106311082839966, |
|
"learning_rate": 3.139717425431711e-07, |
|
"loss": 1.0891, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.9701726844583988, |
|
"grad_norm": 0.29504525661468506, |
|
"learning_rate": 3.0612244897959183e-07, |
|
"loss": 1.0303, |
|
"step": 1236 |
|
}, |
|
{ |
|
"epoch": 0.9709576138147566, |
|
"grad_norm": 0.3173236548900604, |
|
"learning_rate": 2.982731554160126e-07, |
|
"loss": 1.0699, |
|
"step": 1237 |
|
}, |
|
{ |
|
"epoch": 0.9717425431711146, |
|
"grad_norm": 0.2948251962661743, |
|
"learning_rate": 2.904238618524333e-07, |
|
"loss": 1.0273, |
|
"step": 1238 |
|
}, |
|
{ |
|
"epoch": 0.9725274725274725, |
|
"grad_norm": 0.3042560815811157, |
|
"learning_rate": 2.8257456828885403e-07, |
|
"loss": 1.019, |
|
"step": 1239 |
|
}, |
|
{ |
|
"epoch": 0.9733124018838305, |
|
"grad_norm": 0.3046058118343353, |
|
"learning_rate": 2.7472527472527475e-07, |
|
"loss": 1.0301, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.9740973312401884, |
|
"grad_norm": 0.2964264452457428, |
|
"learning_rate": 2.6687598116169546e-07, |
|
"loss": 1.055, |
|
"step": 1241 |
|
}, |
|
{ |
|
"epoch": 0.9748822605965463, |
|
"grad_norm": 0.43890246748924255, |
|
"learning_rate": 2.590266875981162e-07, |
|
"loss": 1.0003, |
|
"step": 1242 |
|
}, |
|
{ |
|
"epoch": 0.9756671899529042, |
|
"grad_norm": 0.3069480359554291, |
|
"learning_rate": 2.5117739403453695e-07, |
|
"loss": 1.0983, |
|
"step": 1243 |
|
}, |
|
{ |
|
"epoch": 0.9764521193092621, |
|
"grad_norm": 0.2868310213088989, |
|
"learning_rate": 2.433281004709576e-07, |
|
"loss": 1.0251, |
|
"step": 1244 |
|
}, |
|
{ |
|
"epoch": 0.9772370486656201, |
|
"grad_norm": 0.28800535202026367, |
|
"learning_rate": 2.3547880690737838e-07, |
|
"loss": 1.0251, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.978021978021978, |
|
"grad_norm": 0.290998637676239, |
|
"learning_rate": 2.2762951334379907e-07, |
|
"loss": 0.996, |
|
"step": 1246 |
|
}, |
|
{ |
|
"epoch": 0.978806907378336, |
|
"grad_norm": 0.3080032467842102, |
|
"learning_rate": 2.197802197802198e-07, |
|
"loss": 1.1012, |
|
"step": 1247 |
|
}, |
|
{ |
|
"epoch": 0.9795918367346939, |
|
"grad_norm": 0.3035239577293396, |
|
"learning_rate": 2.1193092621664052e-07, |
|
"loss": 1.0589, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 0.9803767660910518, |
|
"grad_norm": 0.3100905120372772, |
|
"learning_rate": 2.0408163265306121e-07, |
|
"loss": 1.0276, |
|
"step": 1249 |
|
}, |
|
{ |
|
"epoch": 0.9811616954474097, |
|
"grad_norm": 0.30332428216934204, |
|
"learning_rate": 1.9623233908948196e-07, |
|
"loss": 1.0549, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.9819466248037677, |
|
"grad_norm": 0.2909112572669983, |
|
"learning_rate": 1.8838304552590267e-07, |
|
"loss": 1.0492, |
|
"step": 1251 |
|
}, |
|
{ |
|
"epoch": 0.9827315541601256, |
|
"grad_norm": 0.2939865291118622, |
|
"learning_rate": 1.805337519623234e-07, |
|
"loss": 1.0673, |
|
"step": 1252 |
|
}, |
|
{ |
|
"epoch": 0.9835164835164835, |
|
"grad_norm": 0.31474772095680237, |
|
"learning_rate": 1.726844583987441e-07, |
|
"loss": 1.0801, |
|
"step": 1253 |
|
}, |
|
{ |
|
"epoch": 0.9843014128728415, |
|
"grad_norm": 0.2975396513938904, |
|
"learning_rate": 1.6483516483516484e-07, |
|
"loss": 1.0645, |
|
"step": 1254 |
|
}, |
|
{ |
|
"epoch": 0.9850863422291993, |
|
"grad_norm": 0.2967347204685211, |
|
"learning_rate": 1.5698587127158556e-07, |
|
"loss": 0.9959, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.9858712715855573, |
|
"grad_norm": 0.2978781461715698, |
|
"learning_rate": 1.491365777080063e-07, |
|
"loss": 1.0541, |
|
"step": 1256 |
|
}, |
|
{ |
|
"epoch": 0.9866562009419152, |
|
"grad_norm": 0.37754517793655396, |
|
"learning_rate": 1.4128728414442702e-07, |
|
"loss": 1.0134, |
|
"step": 1257 |
|
}, |
|
{ |
|
"epoch": 0.9874411302982732, |
|
"grad_norm": 0.28837668895721436, |
|
"learning_rate": 1.3343799058084773e-07, |
|
"loss": 1.0435, |
|
"step": 1258 |
|
}, |
|
{ |
|
"epoch": 0.9882260596546311, |
|
"grad_norm": 0.2922952473163605, |
|
"learning_rate": 1.2558869701726847e-07, |
|
"loss": 1.0211, |
|
"step": 1259 |
|
}, |
|
{ |
|
"epoch": 0.989010989010989, |
|
"grad_norm": 0.37126073241233826, |
|
"learning_rate": 1.1773940345368919e-07, |
|
"loss": 1.107, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.9897959183673469, |
|
"grad_norm": 0.30882400274276733, |
|
"learning_rate": 1.098901098901099e-07, |
|
"loss": 1.0274, |
|
"step": 1261 |
|
}, |
|
{ |
|
"epoch": 0.9905808477237049, |
|
"grad_norm": 0.30320990085601807, |
|
"learning_rate": 1.0204081632653061e-07, |
|
"loss": 1.0478, |
|
"step": 1262 |
|
}, |
|
{ |
|
"epoch": 0.9913657770800628, |
|
"grad_norm": 0.305622398853302, |
|
"learning_rate": 9.419152276295134e-08, |
|
"loss": 1.0782, |
|
"step": 1263 |
|
}, |
|
{ |
|
"epoch": 0.9921507064364207, |
|
"grad_norm": 0.2983652353286743, |
|
"learning_rate": 8.634222919937205e-08, |
|
"loss": 1.0486, |
|
"step": 1264 |
|
}, |
|
{ |
|
"epoch": 0.9929356357927787, |
|
"grad_norm": 0.3014610707759857, |
|
"learning_rate": 7.849293563579278e-08, |
|
"loss": 1.0696, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.9937205651491365, |
|
"grad_norm": 0.3930485248565674, |
|
"learning_rate": 7.064364207221351e-08, |
|
"loss": 1.0056, |
|
"step": 1266 |
|
}, |
|
{ |
|
"epoch": 0.9945054945054945, |
|
"grad_norm": 0.30500683188438416, |
|
"learning_rate": 6.279434850863424e-08, |
|
"loss": 1.0317, |
|
"step": 1267 |
|
}, |
|
{ |
|
"epoch": 0.9952904238618524, |
|
"grad_norm": 0.29284876585006714, |
|
"learning_rate": 5.494505494505495e-08, |
|
"loss": 1.0596, |
|
"step": 1268 |
|
}, |
|
{ |
|
"epoch": 0.9960753532182104, |
|
"grad_norm": 0.28736308217048645, |
|
"learning_rate": 4.709576138147567e-08, |
|
"loss": 0.9853, |
|
"step": 1269 |
|
}, |
|
{ |
|
"epoch": 0.9968602825745683, |
|
"grad_norm": 0.2951013445854187, |
|
"learning_rate": 3.924646781789639e-08, |
|
"loss": 1.0667, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.9976452119309263, |
|
"grad_norm": 0.396849662065506, |
|
"learning_rate": 3.139717425431712e-08, |
|
"loss": 1.1164, |
|
"step": 1271 |
|
}, |
|
{ |
|
"epoch": 0.9984301412872841, |
|
"grad_norm": 0.3089428246021271, |
|
"learning_rate": 2.3547880690737834e-08, |
|
"loss": 1.0651, |
|
"step": 1272 |
|
}, |
|
{ |
|
"epoch": 0.9992150706436421, |
|
"grad_norm": 0.295634388923645, |
|
"learning_rate": 1.569858712715856e-08, |
|
"loss": 1.027, |
|
"step": 1273 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.2972462475299835, |
|
"learning_rate": 7.84929356357928e-09, |
|
"loss": 1.0419, |
|
"step": 1274 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 1274, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 0, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.964361379037053e+18, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|