|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.4970273483947683, |
|
"eval_steps": 105, |
|
"global_step": 630, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0023781212841854932, |
|
"grad_norm": 32.74397118558861, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 2.9478, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0023781212841854932, |
|
"eval_loss": 3.373392343521118, |
|
"eval_runtime": 78.9756, |
|
"eval_samples_per_second": 8.091, |
|
"eval_steps_per_second": 1.013, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0047562425683709865, |
|
"grad_norm": 37.080911592721954, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 2.8264, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.007134363852556481, |
|
"grad_norm": 31.107267997702266, |
|
"learning_rate": 1.5e-06, |
|
"loss": 2.9936, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.009512485136741973, |
|
"grad_norm": 27.574905774161167, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 2.8944, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.011890606420927468, |
|
"grad_norm": 13.643957484299273, |
|
"learning_rate": 2.5e-06, |
|
"loss": 2.7687, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.014268727705112961, |
|
"grad_norm": 15.077028980653411, |
|
"learning_rate": 3e-06, |
|
"loss": 2.6623, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.016646848989298454, |
|
"grad_norm": 14.569557474559408, |
|
"learning_rate": 3.5e-06, |
|
"loss": 2.9007, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.019024970273483946, |
|
"grad_norm": 13.894984550517007, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 2.8359, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02140309155766944, |
|
"grad_norm": 13.546442233583257, |
|
"learning_rate": 4.5e-06, |
|
"loss": 2.7583, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.023781212841854936, |
|
"grad_norm": 11.166476039936938, |
|
"learning_rate": 5e-06, |
|
"loss": 2.518, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.026159334126040427, |
|
"grad_norm": 7.787448179397784, |
|
"learning_rate": 5.500000000000001e-06, |
|
"loss": 2.6494, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.028537455410225922, |
|
"grad_norm": 4.72349822440695, |
|
"learning_rate": 6e-06, |
|
"loss": 2.4022, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.030915576694411414, |
|
"grad_norm": 4.100722460414476, |
|
"learning_rate": 6.5000000000000004e-06, |
|
"loss": 2.3933, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.03329369797859691, |
|
"grad_norm": 4.193151112965372, |
|
"learning_rate": 7e-06, |
|
"loss": 2.5468, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0356718192627824, |
|
"grad_norm": 5.502246954578136, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 2.5126, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.03804994054696789, |
|
"grad_norm": 5.716937946349337, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.6761, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.04042806183115339, |
|
"grad_norm": 3.008867017303434, |
|
"learning_rate": 8.5e-06, |
|
"loss": 2.4264, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.04280618311533888, |
|
"grad_norm": 3.6016120293217178, |
|
"learning_rate": 9e-06, |
|
"loss": 2.3836, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.04518430439952437, |
|
"grad_norm": 2.1431386543975908, |
|
"learning_rate": 9.5e-06, |
|
"loss": 2.4879, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.04756242568370987, |
|
"grad_norm": 3.838539096237921, |
|
"learning_rate": 1e-05, |
|
"loss": 2.4116, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04994054696789536, |
|
"grad_norm": 1.9119200890173822, |
|
"learning_rate": 1.0500000000000001e-05, |
|
"loss": 2.6716, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.052318668252080855, |
|
"grad_norm": 2.0853737807318904, |
|
"learning_rate": 1.1000000000000001e-05, |
|
"loss": 2.4523, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.054696789536266346, |
|
"grad_norm": 1.6109539551135528, |
|
"learning_rate": 1.15e-05, |
|
"loss": 2.3086, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.057074910820451845, |
|
"grad_norm": 2.575933824126331, |
|
"learning_rate": 1.2e-05, |
|
"loss": 2.5576, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.059453032104637336, |
|
"grad_norm": 1.4827660467902501, |
|
"learning_rate": 1.25e-05, |
|
"loss": 2.264, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.06183115338882283, |
|
"grad_norm": 2.1022593224903128, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"loss": 2.8781, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.06420927467300833, |
|
"grad_norm": 2.756765764204405, |
|
"learning_rate": 1.3500000000000001e-05, |
|
"loss": 2.5071, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.06658739595719382, |
|
"grad_norm": 1.6249455774563617, |
|
"learning_rate": 1.4e-05, |
|
"loss": 2.5085, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.06896551724137931, |
|
"grad_norm": 1.8653671914008603, |
|
"learning_rate": 1.45e-05, |
|
"loss": 2.4368, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.0713436385255648, |
|
"grad_norm": 1.867632087572305, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 2.1204, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07372175980975029, |
|
"grad_norm": 1.2367329590346625, |
|
"learning_rate": 1.55e-05, |
|
"loss": 2.1118, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.07609988109393578, |
|
"grad_norm": 1.5969207000997636, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 2.4747, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.07847800237812129, |
|
"grad_norm": 1.5075133034220278, |
|
"learning_rate": 1.65e-05, |
|
"loss": 2.3349, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.08085612366230678, |
|
"grad_norm": 1.4803692479229955, |
|
"learning_rate": 1.7e-05, |
|
"loss": 2.3915, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.08323424494649227, |
|
"grad_norm": 2.692949771759104, |
|
"learning_rate": 1.7500000000000002e-05, |
|
"loss": 2.6397, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.08561236623067776, |
|
"grad_norm": 1.3636512770329847, |
|
"learning_rate": 1.8e-05, |
|
"loss": 2.2944, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.08799048751486326, |
|
"grad_norm": 1.8310086696195464, |
|
"learning_rate": 1.8500000000000002e-05, |
|
"loss": 2.2614, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.09036860879904875, |
|
"grad_norm": 1.7765027708264853, |
|
"learning_rate": 1.9e-05, |
|
"loss": 2.3579, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.09274673008323424, |
|
"grad_norm": 1.4484769960901491, |
|
"learning_rate": 1.95e-05, |
|
"loss": 2.4548, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.09512485136741974, |
|
"grad_norm": 3.1520205275209414, |
|
"learning_rate": 2e-05, |
|
"loss": 2.5208, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09750297265160524, |
|
"grad_norm": 1.5897739849482102, |
|
"learning_rate": 1.9999981652287733e-05, |
|
"loss": 2.4216, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.09988109393579073, |
|
"grad_norm": 2.4520591326987495, |
|
"learning_rate": 1.999992660921826e-05, |
|
"loss": 2.2326, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.10225921521997622, |
|
"grad_norm": 1.4817926319377914, |
|
"learning_rate": 1.999983487099356e-05, |
|
"loss": 2.541, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.10463733650416171, |
|
"grad_norm": 2.4022755616863956, |
|
"learning_rate": 1.999970643795027e-05, |
|
"loss": 2.3645, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.1070154577883472, |
|
"grad_norm": 2.476073007712477, |
|
"learning_rate": 1.9999541310559686e-05, |
|
"loss": 2.5051, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.10939357907253269, |
|
"grad_norm": 3.428213096316913, |
|
"learning_rate": 1.9999339489427746e-05, |
|
"loss": 2.3605, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.1117717003567182, |
|
"grad_norm": 2.464783346708793, |
|
"learning_rate": 1.9999100975295046e-05, |
|
"loss": 2.3785, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.11414982164090369, |
|
"grad_norm": 2.0686444585541754, |
|
"learning_rate": 1.999882576903682e-05, |
|
"loss": 2.6035, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.11652794292508918, |
|
"grad_norm": 2.30832669621963, |
|
"learning_rate": 1.9998513871662945e-05, |
|
"loss": 2.3982, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.11890606420927467, |
|
"grad_norm": 1.7443884357427357, |
|
"learning_rate": 1.9998165284317944e-05, |
|
"loss": 2.2344, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.12128418549346016, |
|
"grad_norm": 1.862498116501275, |
|
"learning_rate": 1.999778000828098e-05, |
|
"loss": 2.0631, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.12366230677764566, |
|
"grad_norm": 1.3054707043181313, |
|
"learning_rate": 1.9997358044965833e-05, |
|
"loss": 2.4434, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.12604042806183116, |
|
"grad_norm": 2.6645481874919583, |
|
"learning_rate": 1.9996899395920915e-05, |
|
"loss": 2.4647, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.12841854934601665, |
|
"grad_norm": 1.6884769054479885, |
|
"learning_rate": 1.999640406282926e-05, |
|
"loss": 2.2864, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.13079667063020214, |
|
"grad_norm": 2.281038989939936, |
|
"learning_rate": 1.9995872047508516e-05, |
|
"loss": 2.2999, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.13317479191438764, |
|
"grad_norm": 1.8457967068582515, |
|
"learning_rate": 1.9995303351910934e-05, |
|
"loss": 2.5117, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.13555291319857313, |
|
"grad_norm": 1.593901860141829, |
|
"learning_rate": 1.9994697978123363e-05, |
|
"loss": 2.1696, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.13793103448275862, |
|
"grad_norm": 2.168613470386471, |
|
"learning_rate": 1.9994055928367256e-05, |
|
"loss": 2.6852, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.1403091557669441, |
|
"grad_norm": 1.5398971278043392, |
|
"learning_rate": 1.999337720499863e-05, |
|
"loss": 2.396, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.1426872770511296, |
|
"grad_norm": 1.7618365641290346, |
|
"learning_rate": 1.99926618105081e-05, |
|
"loss": 2.047, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.1450653983353151, |
|
"grad_norm": 1.3935889380673343, |
|
"learning_rate": 1.9991909747520835e-05, |
|
"loss": 2.1071, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.14744351961950058, |
|
"grad_norm": 1.3063537354837544, |
|
"learning_rate": 1.999112101879656e-05, |
|
"loss": 2.3992, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.14982164090368608, |
|
"grad_norm": 1.025531260937785, |
|
"learning_rate": 1.9990295627229544e-05, |
|
"loss": 2.4764, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.15219976218787157, |
|
"grad_norm": 1.0662434476421614, |
|
"learning_rate": 1.99894335758486e-05, |
|
"loss": 2.0608, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.1545778834720571, |
|
"grad_norm": 1.335312659171346, |
|
"learning_rate": 1.9988534867817065e-05, |
|
"loss": 2.0345, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.15695600475624258, |
|
"grad_norm": 1.7866360477276542, |
|
"learning_rate": 1.9987599506432785e-05, |
|
"loss": 2.4781, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.15933412604042807, |
|
"grad_norm": 1.3661388854405736, |
|
"learning_rate": 1.9986627495128105e-05, |
|
"loss": 2.3895, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.16171224732461356, |
|
"grad_norm": 1.69413799763372, |
|
"learning_rate": 1.9985618837469864e-05, |
|
"loss": 2.415, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.16409036860879905, |
|
"grad_norm": 1.6458921632697698, |
|
"learning_rate": 1.998457353715938e-05, |
|
"loss": 2.4017, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.16646848989298454, |
|
"grad_norm": 1.705150979387567, |
|
"learning_rate": 1.998349159803241e-05, |
|
"loss": 2.317, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.16884661117717004, |
|
"grad_norm": 1.4786622132550975, |
|
"learning_rate": 1.9982373024059195e-05, |
|
"loss": 2.4046, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.17122473246135553, |
|
"grad_norm": 1.1257378401253821, |
|
"learning_rate": 1.998121781934438e-05, |
|
"loss": 2.196, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.17360285374554102, |
|
"grad_norm": 1.2450243917414825, |
|
"learning_rate": 1.9980025988127037e-05, |
|
"loss": 2.1235, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1759809750297265, |
|
"grad_norm": 1.2574977788035384, |
|
"learning_rate": 1.9978797534780646e-05, |
|
"loss": 2.4466, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.178359096313912, |
|
"grad_norm": 1.390309850165232, |
|
"learning_rate": 1.9977532463813064e-05, |
|
"loss": 2.4469, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.1807372175980975, |
|
"grad_norm": 1.2146735833909619, |
|
"learning_rate": 1.9976230779866527e-05, |
|
"loss": 2.3705, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.18311533888228299, |
|
"grad_norm": 1.26356031715395, |
|
"learning_rate": 1.9974892487717613e-05, |
|
"loss": 2.4926, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.18549346016646848, |
|
"grad_norm": 1.3934309027656608, |
|
"learning_rate": 1.997351759227725e-05, |
|
"loss": 2.47, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.187871581450654, |
|
"grad_norm": 1.129998342751621, |
|
"learning_rate": 1.9972106098590665e-05, |
|
"loss": 2.3718, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.1902497027348395, |
|
"grad_norm": 1.3957720808228478, |
|
"learning_rate": 1.9970658011837404e-05, |
|
"loss": 2.2057, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.19262782401902498, |
|
"grad_norm": 1.5835508346410572, |
|
"learning_rate": 1.9969173337331283e-05, |
|
"loss": 2.2551, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.19500594530321047, |
|
"grad_norm": 1.0726363834452401, |
|
"learning_rate": 1.996765208052037e-05, |
|
"loss": 1.9962, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.19738406658739596, |
|
"grad_norm": 1.5380248440103288, |
|
"learning_rate": 1.9966094246986983e-05, |
|
"loss": 2.1986, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.19976218787158145, |
|
"grad_norm": 1.2953387209833067, |
|
"learning_rate": 1.9964499842447665e-05, |
|
"loss": 2.5842, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.20214030915576695, |
|
"grad_norm": 1.0712033116668103, |
|
"learning_rate": 1.9962868872753144e-05, |
|
"loss": 2.1298, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.20451843043995244, |
|
"grad_norm": 1.095761902776689, |
|
"learning_rate": 1.996120134388834e-05, |
|
"loss": 2.3641, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.20689655172413793, |
|
"grad_norm": 1.186679631328553, |
|
"learning_rate": 1.995949726197231e-05, |
|
"loss": 2.4801, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.20927467300832342, |
|
"grad_norm": 1.1887498108170933, |
|
"learning_rate": 1.9957756633258264e-05, |
|
"loss": 2.2866, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.2116527942925089, |
|
"grad_norm": 1.0909023774872124, |
|
"learning_rate": 1.9955979464133515e-05, |
|
"loss": 2.2916, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.2140309155766944, |
|
"grad_norm": 1.1077175360558418, |
|
"learning_rate": 1.995416576111945e-05, |
|
"loss": 2.1077, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.2164090368608799, |
|
"grad_norm": 1.1526064687436712, |
|
"learning_rate": 1.9952315530871537e-05, |
|
"loss": 2.2723, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.21878715814506539, |
|
"grad_norm": 1.1818210038912647, |
|
"learning_rate": 1.9950428780179274e-05, |
|
"loss": 2.2338, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.2211652794292509, |
|
"grad_norm": 1.164942154271255, |
|
"learning_rate": 1.994850551596617e-05, |
|
"loss": 2.3817, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.2235434007134364, |
|
"grad_norm": 1.339398993177121, |
|
"learning_rate": 1.9946545745289727e-05, |
|
"loss": 2.5508, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.2259215219976219, |
|
"grad_norm": 1.3267763052855093, |
|
"learning_rate": 1.9944549475341404e-05, |
|
"loss": 2.247, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.22829964328180738, |
|
"grad_norm": 1.1953250811556597, |
|
"learning_rate": 1.99425167134466e-05, |
|
"loss": 2.3373, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.23067776456599287, |
|
"grad_norm": 1.4321452409301854, |
|
"learning_rate": 1.9940447467064624e-05, |
|
"loss": 2.4776, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.23305588585017836, |
|
"grad_norm": 1.0224444212683161, |
|
"learning_rate": 1.9938341743788658e-05, |
|
"loss": 2.1837, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.23543400713436385, |
|
"grad_norm": 1.0977574950238398, |
|
"learning_rate": 1.9936199551345744e-05, |
|
"loss": 2.1478, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.23781212841854935, |
|
"grad_norm": 1.0660069054078747, |
|
"learning_rate": 1.9934020897596752e-05, |
|
"loss": 2.2816, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.24019024970273484, |
|
"grad_norm": 1.0312993672336248, |
|
"learning_rate": 1.9931805790536342e-05, |
|
"loss": 2.2468, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.24256837098692033, |
|
"grad_norm": 1.1278898252066067, |
|
"learning_rate": 1.9929554238292944e-05, |
|
"loss": 2.0526, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.24494649227110582, |
|
"grad_norm": 1.288343002944789, |
|
"learning_rate": 1.992726624912872e-05, |
|
"loss": 2.512, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.2473246135552913, |
|
"grad_norm": 1.1840020386119305, |
|
"learning_rate": 1.992494183143955e-05, |
|
"loss": 2.6356, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.2497027348394768, |
|
"grad_norm": 1.1014678408276726, |
|
"learning_rate": 1.9922580993754985e-05, |
|
"loss": 2.3521, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.2497027348394768, |
|
"eval_loss": 2.5365779399871826, |
|
"eval_runtime": 66.0796, |
|
"eval_samples_per_second": 9.67, |
|
"eval_steps_per_second": 1.211, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.2520808561236623, |
|
"grad_norm": 0.9957844757920508, |
|
"learning_rate": 1.9920183744738208e-05, |
|
"loss": 2.355, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.2544589774078478, |
|
"grad_norm": 1.0070598447313825, |
|
"learning_rate": 1.9917750093186036e-05, |
|
"loss": 2.1747, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.2568370986920333, |
|
"grad_norm": 1.2463453868295562, |
|
"learning_rate": 1.9915280048028853e-05, |
|
"loss": 2.4131, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.25921521997621877, |
|
"grad_norm": 1.158558292534161, |
|
"learning_rate": 1.9912773618330595e-05, |
|
"loss": 2.4527, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.2615933412604043, |
|
"grad_norm": 1.1875643459332377, |
|
"learning_rate": 1.9910230813288713e-05, |
|
"loss": 2.1523, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.26397146254458975, |
|
"grad_norm": 0.892269173897758, |
|
"learning_rate": 1.9907651642234138e-05, |
|
"loss": 1.9606, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.26634958382877527, |
|
"grad_norm": 1.181952902180908, |
|
"learning_rate": 1.9905036114631247e-05, |
|
"loss": 2.3201, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.26872770511296074, |
|
"grad_norm": 0.9689153704257877, |
|
"learning_rate": 1.990238424007783e-05, |
|
"loss": 2.2329, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.27110582639714625, |
|
"grad_norm": 1.3665918769424286, |
|
"learning_rate": 1.989969602830505e-05, |
|
"loss": 2.2387, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.2734839476813318, |
|
"grad_norm": 1.0478434719151144, |
|
"learning_rate": 1.9896971489177417e-05, |
|
"loss": 2.2798, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.27586206896551724, |
|
"grad_norm": 1.5752154316391798, |
|
"learning_rate": 1.9894210632692745e-05, |
|
"loss": 2.201, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.27824019024970276, |
|
"grad_norm": 1.0264277011384757, |
|
"learning_rate": 1.9891413468982112e-05, |
|
"loss": 2.2756, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.2806183115338882, |
|
"grad_norm": 1.1750703393359614, |
|
"learning_rate": 1.988858000830983e-05, |
|
"loss": 2.1907, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.28299643281807374, |
|
"grad_norm": 0.9456957190962577, |
|
"learning_rate": 1.9885710261073402e-05, |
|
"loss": 2.2993, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.2853745541022592, |
|
"grad_norm": 1.37591692336223, |
|
"learning_rate": 1.9882804237803487e-05, |
|
"loss": 2.0751, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.2877526753864447, |
|
"grad_norm": 0.991181263305241, |
|
"learning_rate": 1.9879861949163863e-05, |
|
"loss": 2.1946, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.2901307966706302, |
|
"grad_norm": 1.2826616603092615, |
|
"learning_rate": 1.9876883405951378e-05, |
|
"loss": 2.3084, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.2925089179548157, |
|
"grad_norm": 1.3162982027829009, |
|
"learning_rate": 1.987386861909593e-05, |
|
"loss": 2.294, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.29488703923900117, |
|
"grad_norm": 1.086311999313279, |
|
"learning_rate": 1.98708175996604e-05, |
|
"loss": 2.3025, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.2972651605231867, |
|
"grad_norm": 1.10683170372015, |
|
"learning_rate": 1.986773035884064e-05, |
|
"loss": 2.3447, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.29964328180737215, |
|
"grad_norm": 1.090568761480393, |
|
"learning_rate": 1.9864606907965407e-05, |
|
"loss": 2.4104, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.30202140309155767, |
|
"grad_norm": 1.4024759238343605, |
|
"learning_rate": 1.986144725849634e-05, |
|
"loss": 2.298, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.30439952437574314, |
|
"grad_norm": 0.9324914520062791, |
|
"learning_rate": 1.9858251422027903e-05, |
|
"loss": 2.1123, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.30677764565992865, |
|
"grad_norm": 1.3818136151492852, |
|
"learning_rate": 1.9855019410287355e-05, |
|
"loss": 2.2786, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.3091557669441142, |
|
"grad_norm": 0.9879756737720099, |
|
"learning_rate": 1.98517512351347e-05, |
|
"loss": 2.2735, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.31153388822829964, |
|
"grad_norm": 1.4107106057474024, |
|
"learning_rate": 1.9848446908562647e-05, |
|
"loss": 2.2421, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.31391200951248516, |
|
"grad_norm": 0.978862094447652, |
|
"learning_rate": 1.9845106442696563e-05, |
|
"loss": 2.4152, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.3162901307966706, |
|
"grad_norm": 1.3714074038447606, |
|
"learning_rate": 1.9841729849794427e-05, |
|
"loss": 2.4567, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.31866825208085614, |
|
"grad_norm": 1.030641093673837, |
|
"learning_rate": 1.983831714224679e-05, |
|
"loss": 2.3015, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.3210463733650416, |
|
"grad_norm": 1.1744699755999302, |
|
"learning_rate": 1.9834868332576727e-05, |
|
"loss": 2.2878, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.3234244946492271, |
|
"grad_norm": 0.9733999816490441, |
|
"learning_rate": 1.9831383433439798e-05, |
|
"loss": 2.1571, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.3258026159334126, |
|
"grad_norm": 1.0470367999253474, |
|
"learning_rate": 1.982786245762398e-05, |
|
"loss": 2.0943, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.3281807372175981, |
|
"grad_norm": 1.0748276455064096, |
|
"learning_rate": 1.9824305418049645e-05, |
|
"loss": 2.4156, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.33055885850178357, |
|
"grad_norm": 1.0220509349947084, |
|
"learning_rate": 1.9820712327769503e-05, |
|
"loss": 2.1898, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.3329369797859691, |
|
"grad_norm": 0.9811166423920332, |
|
"learning_rate": 1.9817083199968552e-05, |
|
"loss": 2.3449, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.33531510107015455, |
|
"grad_norm": 1.0664757695722766, |
|
"learning_rate": 1.9813418047964025e-05, |
|
"loss": 2.1514, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.3376932223543401, |
|
"grad_norm": 1.1228830278366924, |
|
"learning_rate": 1.9809716885205363e-05, |
|
"loss": 2.3371, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.3400713436385256, |
|
"grad_norm": 1.0703957613617774, |
|
"learning_rate": 1.980597972527413e-05, |
|
"loss": 2.2577, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.34244946492271106, |
|
"grad_norm": 0.9971842999532138, |
|
"learning_rate": 1.9802206581883992e-05, |
|
"loss": 2.2048, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.3448275862068966, |
|
"grad_norm": 0.9969712850303254, |
|
"learning_rate": 1.979839746888067e-05, |
|
"loss": 2.1725, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.34720570749108204, |
|
"grad_norm": 0.9782490093980141, |
|
"learning_rate": 1.979455240024186e-05, |
|
"loss": 2.1598, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.34958382877526756, |
|
"grad_norm": 1.1595035293528873, |
|
"learning_rate": 1.97906713900772e-05, |
|
"loss": 2.1812, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.351961950059453, |
|
"grad_norm": 1.0488323565717943, |
|
"learning_rate": 1.9786754452628226e-05, |
|
"loss": 2.126, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.35434007134363854, |
|
"grad_norm": 1.0236205683546673, |
|
"learning_rate": 1.9782801602268306e-05, |
|
"loss": 1.9399, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.356718192627824, |
|
"grad_norm": 0.983049547537296, |
|
"learning_rate": 1.9778812853502592e-05, |
|
"loss": 2.0336, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.3590963139120095, |
|
"grad_norm": 1.0856474713800959, |
|
"learning_rate": 1.9774788220967968e-05, |
|
"loss": 2.2103, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.361474435196195, |
|
"grad_norm": 1.098143269144179, |
|
"learning_rate": 1.9770727719432994e-05, |
|
"loss": 2.1425, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.3638525564803805, |
|
"grad_norm": 1.1908904777112574, |
|
"learning_rate": 1.9766631363797852e-05, |
|
"loss": 2.2516, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.36623067776456597, |
|
"grad_norm": 1.1823343263781934, |
|
"learning_rate": 1.9762499169094288e-05, |
|
"loss": 2.0991, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.3686087990487515, |
|
"grad_norm": 1.1543274307271654, |
|
"learning_rate": 1.9758331150485576e-05, |
|
"loss": 2.2917, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.37098692033293695, |
|
"grad_norm": 1.1828452156246019, |
|
"learning_rate": 1.9754127323266426e-05, |
|
"loss": 2.3577, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.3733650416171225, |
|
"grad_norm": 1.2458434785978698, |
|
"learning_rate": 1.9749887702862972e-05, |
|
"loss": 2.2291, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.375743162901308, |
|
"grad_norm": 1.0632348458757013, |
|
"learning_rate": 1.9745612304832672e-05, |
|
"loss": 2.495, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.37812128418549346, |
|
"grad_norm": 1.2413557275846534, |
|
"learning_rate": 1.9741301144864284e-05, |
|
"loss": 2.3006, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.380499405469679, |
|
"grad_norm": 1.068837985332943, |
|
"learning_rate": 1.9736954238777793e-05, |
|
"loss": 2.2228, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.38287752675386444, |
|
"grad_norm": 1.181973772137545, |
|
"learning_rate": 1.9732571602524353e-05, |
|
"loss": 2.3419, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.38525564803804996, |
|
"grad_norm": 0.9361759344356807, |
|
"learning_rate": 1.972815325218624e-05, |
|
"loss": 2.2727, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.3876337693222354, |
|
"grad_norm": 1.2300672941710984, |
|
"learning_rate": 1.9723699203976768e-05, |
|
"loss": 2.3947, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.39001189060642094, |
|
"grad_norm": 0.9647921025871186, |
|
"learning_rate": 1.9719209474240263e-05, |
|
"loss": 1.8388, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.3923900118906064, |
|
"grad_norm": 1.1390311715526416, |
|
"learning_rate": 1.971468407945198e-05, |
|
"loss": 2.4054, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.3947681331747919, |
|
"grad_norm": 0.9830051867519547, |
|
"learning_rate": 1.9710123036218044e-05, |
|
"loss": 2.0355, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.3971462544589774, |
|
"grad_norm": 1.1244517585073737, |
|
"learning_rate": 1.97055263612754e-05, |
|
"loss": 2.0188, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.3995243757431629, |
|
"grad_norm": 1.0256020852263494, |
|
"learning_rate": 1.9700894071491736e-05, |
|
"loss": 2.0774, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.40190249702734837, |
|
"grad_norm": 1.011023720252716, |
|
"learning_rate": 1.9696226183865436e-05, |
|
"loss": 2.2592, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.4042806183115339, |
|
"grad_norm": 1.046975898884085, |
|
"learning_rate": 1.969152271552552e-05, |
|
"loss": 2.1791, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.40665873959571935, |
|
"grad_norm": 1.1800984480399852, |
|
"learning_rate": 1.9686783683731557e-05, |
|
"loss": 2.3941, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.4090368608799049, |
|
"grad_norm": 1.2459882622321672, |
|
"learning_rate": 1.9682009105873633e-05, |
|
"loss": 2.1522, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.4114149821640904, |
|
"grad_norm": 1.0732133381850257, |
|
"learning_rate": 1.9677198999472257e-05, |
|
"loss": 2.1233, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.41379310344827586, |
|
"grad_norm": 1.2405484917580802, |
|
"learning_rate": 1.967235338217832e-05, |
|
"loss": 2.3016, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.4161712247324614, |
|
"grad_norm": 1.0759940201219593, |
|
"learning_rate": 1.9667472271773026e-05, |
|
"loss": 2.2947, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.41854934601664684, |
|
"grad_norm": 1.2008734320661734, |
|
"learning_rate": 1.9662555686167808e-05, |
|
"loss": 2.2155, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.42092746730083236, |
|
"grad_norm": 0.9303619935178572, |
|
"learning_rate": 1.965760364340429e-05, |
|
"loss": 2.1234, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.4233055885850178, |
|
"grad_norm": 1.3884826767438652, |
|
"learning_rate": 1.9652616161654204e-05, |
|
"loss": 2.2539, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.42568370986920334, |
|
"grad_norm": 0.9947187673832885, |
|
"learning_rate": 1.9647593259219328e-05, |
|
"loss": 2.2052, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.4280618311533888, |
|
"grad_norm": 1.4655922792083054, |
|
"learning_rate": 1.964253495453141e-05, |
|
"loss": 2.1552, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.4304399524375743, |
|
"grad_norm": 1.1481294188693778, |
|
"learning_rate": 1.963744126615212e-05, |
|
"loss": 2.3942, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.4328180737217598, |
|
"grad_norm": 1.239760521409481, |
|
"learning_rate": 1.9632312212772956e-05, |
|
"loss": 2.3091, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.4351961950059453, |
|
"grad_norm": 1.0524654460411744, |
|
"learning_rate": 1.9627147813215207e-05, |
|
"loss": 2.302, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.43757431629013077, |
|
"grad_norm": 1.0231645108607732, |
|
"learning_rate": 1.9621948086429847e-05, |
|
"loss": 2.2334, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.4399524375743163, |
|
"grad_norm": 1.0600582051447691, |
|
"learning_rate": 1.9616713051497496e-05, |
|
"loss": 2.2044, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.4423305588585018, |
|
"grad_norm": 1.0861978175484295, |
|
"learning_rate": 1.9611442727628344e-05, |
|
"loss": 2.3267, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.4447086801426873, |
|
"grad_norm": 1.0122924353396487, |
|
"learning_rate": 1.960613713416206e-05, |
|
"loss": 2.2327, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.4470868014268728, |
|
"grad_norm": 1.1275635495135592, |
|
"learning_rate": 1.9600796290567747e-05, |
|
"loss": 2.2474, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.44946492271105826, |
|
"grad_norm": 1.0778906611663819, |
|
"learning_rate": 1.9595420216443864e-05, |
|
"loss": 2.2777, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.4518430439952438, |
|
"grad_norm": 1.0593499669893551, |
|
"learning_rate": 1.9590008931518133e-05, |
|
"loss": 2.4937, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.45422116527942924, |
|
"grad_norm": 1.0887914371115388, |
|
"learning_rate": 1.9584562455647494e-05, |
|
"loss": 2.2577, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.45659928656361476, |
|
"grad_norm": 1.0280779311785984, |
|
"learning_rate": 1.9579080808818035e-05, |
|
"loss": 2.2352, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.4589774078478002, |
|
"grad_norm": 1.1201705856067985, |
|
"learning_rate": 1.9573564011144873e-05, |
|
"loss": 2.1482, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.46135552913198574, |
|
"grad_norm": 1.0039435227655624, |
|
"learning_rate": 1.9568012082872148e-05, |
|
"loss": 2.1069, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.4637336504161712, |
|
"grad_norm": 1.0523831000821406, |
|
"learning_rate": 1.9562425044372884e-05, |
|
"loss": 1.9268, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.4661117717003567, |
|
"grad_norm": 1.0635880350342213, |
|
"learning_rate": 1.9556802916148963e-05, |
|
"loss": 2.2722, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.4684898929845422, |
|
"grad_norm": 2.4351848601787287, |
|
"learning_rate": 1.955114571883102e-05, |
|
"loss": 2.1402, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.4708680142687277, |
|
"grad_norm": 1.2199308274597462, |
|
"learning_rate": 1.9545453473178384e-05, |
|
"loss": 2.2599, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.47324613555291317, |
|
"grad_norm": 0.9936114796299212, |
|
"learning_rate": 1.9539726200078987e-05, |
|
"loss": 2.0662, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.4756242568370987, |
|
"grad_norm": 1.0692703333507547, |
|
"learning_rate": 1.9533963920549307e-05, |
|
"loss": 2.3739, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4780023781212842, |
|
"grad_norm": 1.0406002686664542, |
|
"learning_rate": 1.9528166655734267e-05, |
|
"loss": 2.3611, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.4803804994054697, |
|
"grad_norm": 1.9375905536343168, |
|
"learning_rate": 1.9522334426907185e-05, |
|
"loss": 2.0971, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.4827586206896552, |
|
"grad_norm": 1.024548704059581, |
|
"learning_rate": 1.951646725546966e-05, |
|
"loss": 2.2498, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.48513674197384066, |
|
"grad_norm": 1.0033895284405978, |
|
"learning_rate": 1.9510565162951538e-05, |
|
"loss": 2.299, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.4875148632580262, |
|
"grad_norm": 1.541631519071697, |
|
"learning_rate": 1.950462817101079e-05, |
|
"loss": 2.4076, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.48989298454221164, |
|
"grad_norm": 0.9499702987331401, |
|
"learning_rate": 1.9498656301433466e-05, |
|
"loss": 2.0754, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.49227110582639716, |
|
"grad_norm": 1.099383371761328, |
|
"learning_rate": 1.9492649576133594e-05, |
|
"loss": 2.2514, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.4946492271105826, |
|
"grad_norm": 0.9296431838496088, |
|
"learning_rate": 1.94866080171531e-05, |
|
"loss": 2.2308, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.49702734839476814, |
|
"grad_norm": 4.140796209905845, |
|
"learning_rate": 1.9480531646661753e-05, |
|
"loss": 2.4388, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.4994054696789536, |
|
"grad_norm": 1.011142238194789, |
|
"learning_rate": 1.9474420486957045e-05, |
|
"loss": 2.2414, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.4994054696789536, |
|
"eval_loss": 2.49302339553833, |
|
"eval_runtime": 65.8636, |
|
"eval_samples_per_second": 9.702, |
|
"eval_steps_per_second": 1.215, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.5017835909631391, |
|
"grad_norm": 0.9941401750604694, |
|
"learning_rate": 1.9468274560464134e-05, |
|
"loss": 2.2182, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.5041617122473246, |
|
"grad_norm": 1.035068690961865, |
|
"learning_rate": 1.9462093889735766e-05, |
|
"loss": 2.3569, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.5065398335315101, |
|
"grad_norm": 1.5052185888965133, |
|
"learning_rate": 1.945587849745217e-05, |
|
"loss": 2.4474, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.5089179548156956, |
|
"grad_norm": 1.01730612362564, |
|
"learning_rate": 1.944962840642099e-05, |
|
"loss": 2.3912, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.5112960760998811, |
|
"grad_norm": 0.9513743221196854, |
|
"learning_rate": 1.9443343639577206e-05, |
|
"loss": 2.2842, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.5136741973840666, |
|
"grad_norm": 0.9759286606572132, |
|
"learning_rate": 1.943702421998303e-05, |
|
"loss": 2.3117, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.5160523186682521, |
|
"grad_norm": 0.9357816333722543, |
|
"learning_rate": 1.9430670170827844e-05, |
|
"loss": 2.1091, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.5184304399524375, |
|
"grad_norm": 1.1804547022456764, |
|
"learning_rate": 1.94242815154281e-05, |
|
"loss": 2.2333, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.5208085612366231, |
|
"grad_norm": 1.0289463651234612, |
|
"learning_rate": 1.9417858277227244e-05, |
|
"loss": 2.0495, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.5231866825208086, |
|
"grad_norm": 0.9297621773216459, |
|
"learning_rate": 1.9411400479795618e-05, |
|
"loss": 2.2102, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.525564803804994, |
|
"grad_norm": 0.9468558483075745, |
|
"learning_rate": 1.9404908146830383e-05, |
|
"loss": 2.2593, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.5279429250891795, |
|
"grad_norm": 1.03418169808611, |
|
"learning_rate": 1.9398381302155435e-05, |
|
"loss": 2.185, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.5303210463733651, |
|
"grad_norm": 0.84914434823741, |
|
"learning_rate": 1.93918199697213e-05, |
|
"loss": 1.8499, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.5326991676575505, |
|
"grad_norm": 1.0824133835143805, |
|
"learning_rate": 1.9385224173605072e-05, |
|
"loss": 2.2171, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.535077288941736, |
|
"grad_norm": 0.9478237708740055, |
|
"learning_rate": 1.9378593938010302e-05, |
|
"loss": 2.2397, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.5374554102259215, |
|
"grad_norm": 2.4277055664527065, |
|
"learning_rate": 1.937192928726692e-05, |
|
"loss": 2.2262, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.539833531510107, |
|
"grad_norm": 1.5223794932859396, |
|
"learning_rate": 1.936523024583115e-05, |
|
"loss": 2.2664, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.5422116527942925, |
|
"grad_norm": 1.2991334808397113, |
|
"learning_rate": 1.9358496838285408e-05, |
|
"loss": 2.1967, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.544589774078478, |
|
"grad_norm": 1.132823106642245, |
|
"learning_rate": 1.9351729089338214e-05, |
|
"loss": 2.2516, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.5469678953626635, |
|
"grad_norm": 1.0872118899678849, |
|
"learning_rate": 1.9344927023824112e-05, |
|
"loss": 2.3602, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.549346016646849, |
|
"grad_norm": 0.9596136441748024, |
|
"learning_rate": 1.933809066670357e-05, |
|
"loss": 2.3944, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.5517241379310345, |
|
"grad_norm": 1.3340981902899025, |
|
"learning_rate": 1.9331220043062894e-05, |
|
"loss": 1.8085, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.5541022592152199, |
|
"grad_norm": 0.9780209403031624, |
|
"learning_rate": 1.9324315178114127e-05, |
|
"loss": 2.2626, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.5564803804994055, |
|
"grad_norm": 0.8441113781121591, |
|
"learning_rate": 1.9317376097194964e-05, |
|
"loss": 1.904, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.558858501783591, |
|
"grad_norm": 0.9000708142025062, |
|
"learning_rate": 1.9310402825768655e-05, |
|
"loss": 2.1239, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.5612366230677764, |
|
"grad_norm": 0.9819927724537073, |
|
"learning_rate": 1.9303395389423918e-05, |
|
"loss": 2.2526, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.5636147443519619, |
|
"grad_norm": 0.9835396213176555, |
|
"learning_rate": 1.9296353813874838e-05, |
|
"loss": 2.1293, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.5659928656361475, |
|
"grad_norm": 0.8705520689865173, |
|
"learning_rate": 1.9289278124960777e-05, |
|
"loss": 1.9911, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.5683709869203329, |
|
"grad_norm": 1.0080028925541162, |
|
"learning_rate": 1.9282168348646268e-05, |
|
"loss": 2.0903, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.5707491082045184, |
|
"grad_norm": 1.0304202570689127, |
|
"learning_rate": 1.927502451102095e-05, |
|
"loss": 2.0784, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.5731272294887039, |
|
"grad_norm": 1.0836386452615874, |
|
"learning_rate": 1.926784663829943e-05, |
|
"loss": 2.2297, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.5755053507728894, |
|
"grad_norm": 0.972327672270083, |
|
"learning_rate": 1.926063475682121e-05, |
|
"loss": 2.2518, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.5778834720570749, |
|
"grad_norm": 1.1250553185490555, |
|
"learning_rate": 1.9253388893050612e-05, |
|
"loss": 2.2419, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.5802615933412604, |
|
"grad_norm": 1.6628057931903235, |
|
"learning_rate": 1.924610907357663e-05, |
|
"loss": 2.2409, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.582639714625446, |
|
"grad_norm": 1.3164993193299928, |
|
"learning_rate": 1.9238795325112867e-05, |
|
"loss": 2.4277, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.5850178359096314, |
|
"grad_norm": 0.9607216662774912, |
|
"learning_rate": 1.9231447674497444e-05, |
|
"loss": 2.0357, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.5873959571938169, |
|
"grad_norm": 0.9753564233974246, |
|
"learning_rate": 1.922406614869287e-05, |
|
"loss": 2.1743, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.5897740784780023, |
|
"grad_norm": 0.9514731158004311, |
|
"learning_rate": 1.9216650774785975e-05, |
|
"loss": 2.2775, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.5921521997621879, |
|
"grad_norm": 1.1514073630104866, |
|
"learning_rate": 1.9209201579987777e-05, |
|
"loss": 2.0307, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.5945303210463734, |
|
"grad_norm": 1.0070268666833808, |
|
"learning_rate": 1.9201718591633417e-05, |
|
"loss": 2.405, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.5969084423305588, |
|
"grad_norm": 0.9724192667298419, |
|
"learning_rate": 1.9194201837182045e-05, |
|
"loss": 2.3375, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.5992865636147443, |
|
"grad_norm": 0.9184225452723915, |
|
"learning_rate": 1.9186651344216703e-05, |
|
"loss": 2.0576, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.6016646848989299, |
|
"grad_norm": 0.9277770597042948, |
|
"learning_rate": 1.9179067140444246e-05, |
|
"loss": 2.065, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.6040428061831153, |
|
"grad_norm": 0.9971241617303254, |
|
"learning_rate": 1.9171449253695233e-05, |
|
"loss": 2.0388, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.6064209274673008, |
|
"grad_norm": 0.968526625269074, |
|
"learning_rate": 1.9163797711923822e-05, |
|
"loss": 2.2154, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.6087990487514863, |
|
"grad_norm": 1.0974259939522593, |
|
"learning_rate": 1.9156112543207674e-05, |
|
"loss": 2.0897, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.6111771700356718, |
|
"grad_norm": 1.0661979784688387, |
|
"learning_rate": 1.9148393775747842e-05, |
|
"loss": 2.3986, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.6135552913198573, |
|
"grad_norm": 1.3741439020152701, |
|
"learning_rate": 1.9140641437868664e-05, |
|
"loss": 2.1249, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.6159334126040428, |
|
"grad_norm": 0.9207355567454468, |
|
"learning_rate": 1.913285555801768e-05, |
|
"loss": 2.2715, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.6183115338882283, |
|
"grad_norm": 1.0171260467381849, |
|
"learning_rate": 1.9125036164765502e-05, |
|
"loss": 2.2638, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.6206896551724138, |
|
"grad_norm": 0.9849622761444283, |
|
"learning_rate": 1.9117183286805726e-05, |
|
"loss": 2.148, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.6230677764565993, |
|
"grad_norm": 1.384730629468189, |
|
"learning_rate": 1.9109296952954826e-05, |
|
"loss": 2.2442, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.6254458977407847, |
|
"grad_norm": 1.066860211861379, |
|
"learning_rate": 1.9101377192152033e-05, |
|
"loss": 2.451, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.6278240190249703, |
|
"grad_norm": 1.1537333721148562, |
|
"learning_rate": 1.909342403345925e-05, |
|
"loss": 2.3076, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.6302021403091558, |
|
"grad_norm": 1.388798426336895, |
|
"learning_rate": 1.9085437506060925e-05, |
|
"loss": 2.3458, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.6325802615933412, |
|
"grad_norm": 1.057136103530956, |
|
"learning_rate": 1.9077417639263966e-05, |
|
"loss": 2.1901, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.6349583828775267, |
|
"grad_norm": 1.0267855180849, |
|
"learning_rate": 1.906936446249761e-05, |
|
"loss": 2.2597, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.6373365041617123, |
|
"grad_norm": 0.984822259782258, |
|
"learning_rate": 1.906127800531333e-05, |
|
"loss": 1.9602, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.6397146254458977, |
|
"grad_norm": 1.0914462759031294, |
|
"learning_rate": 1.905315829738473e-05, |
|
"loss": 2.4246, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.6420927467300832, |
|
"grad_norm": 1.007066333419377, |
|
"learning_rate": 1.9045005368507418e-05, |
|
"loss": 2.2638, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.6444708680142688, |
|
"grad_norm": 0.8517882887076221, |
|
"learning_rate": 1.9036819248598914e-05, |
|
"loss": 1.9905, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.6468489892984542, |
|
"grad_norm": 1.0217039586834744, |
|
"learning_rate": 1.9028599967698533e-05, |
|
"loss": 2.2177, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.6492271105826397, |
|
"grad_norm": 0.9930930260750681, |
|
"learning_rate": 1.902034755596727e-05, |
|
"loss": 2.093, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.6516052318668252, |
|
"grad_norm": 1.4476621256505835, |
|
"learning_rate": 1.9012062043687713e-05, |
|
"loss": 2.2535, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.6539833531510107, |
|
"grad_norm": 0.9679973366177126, |
|
"learning_rate": 1.9003743461263887e-05, |
|
"loss": 2.2304, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.6563614744351962, |
|
"grad_norm": 1.0576995279849084, |
|
"learning_rate": 1.899539183922119e-05, |
|
"loss": 2.0678, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.6587395957193817, |
|
"grad_norm": 0.9196543804308106, |
|
"learning_rate": 1.8987007208206254e-05, |
|
"loss": 2.1106, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.6611177170035671, |
|
"grad_norm": 1.0412924135468502, |
|
"learning_rate": 1.8978589598986838e-05, |
|
"loss": 2.3659, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.6634958382877527, |
|
"grad_norm": 1.1221322457101093, |
|
"learning_rate": 1.8970139042451712e-05, |
|
"loss": 2.0853, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.6658739595719382, |
|
"grad_norm": 0.8927332524290843, |
|
"learning_rate": 1.8961655569610557e-05, |
|
"loss": 2.2388, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.6682520808561236, |
|
"grad_norm": 1.0304293869359518, |
|
"learning_rate": 1.8953139211593838e-05, |
|
"loss": 2.381, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.6706302021403091, |
|
"grad_norm": 1.0508469543032843, |
|
"learning_rate": 1.8944589999652687e-05, |
|
"loss": 2.214, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.6730083234244947, |
|
"grad_norm": 0.8545452252368634, |
|
"learning_rate": 1.8936007965158806e-05, |
|
"loss": 2.0308, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.6753864447086801, |
|
"grad_norm": 1.3101079437708238, |
|
"learning_rate": 1.8927393139604327e-05, |
|
"loss": 2.3389, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.6777645659928656, |
|
"grad_norm": 0.8257605335721794, |
|
"learning_rate": 1.8918745554601726e-05, |
|
"loss": 1.9511, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.6801426872770512, |
|
"grad_norm": 0.9918391764529991, |
|
"learning_rate": 1.891006524188368e-05, |
|
"loss": 2.3694, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.6825208085612366, |
|
"grad_norm": 0.8841124075255721, |
|
"learning_rate": 1.8901352233302964e-05, |
|
"loss": 2.1977, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.6848989298454221, |
|
"grad_norm": 0.8100832981740426, |
|
"learning_rate": 1.8892606560832335e-05, |
|
"loss": 1.9161, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.6872770511296076, |
|
"grad_norm": 0.8738803349826288, |
|
"learning_rate": 1.8883828256564413e-05, |
|
"loss": 2.2419, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.6896551724137931, |
|
"grad_norm": 0.8961748350433363, |
|
"learning_rate": 1.8875017352711547e-05, |
|
"loss": 2.1639, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.6920332936979786, |
|
"grad_norm": 1.0902732587802257, |
|
"learning_rate": 1.886617388160573e-05, |
|
"loss": 2.4486, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.6944114149821641, |
|
"grad_norm": 0.9653296144522788, |
|
"learning_rate": 1.8857297875698455e-05, |
|
"loss": 2.0904, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.6967895362663495, |
|
"grad_norm": 1.0898489125890396, |
|
"learning_rate": 1.8848389367560604e-05, |
|
"loss": 2.2726, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.6991676575505351, |
|
"grad_norm": 0.8892982592369411, |
|
"learning_rate": 1.883944838988232e-05, |
|
"loss": 1.8982, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.7015457788347206, |
|
"grad_norm": 0.9481426246994946, |
|
"learning_rate": 1.8830474975472904e-05, |
|
"loss": 2.1466, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.703923900118906, |
|
"grad_norm": 0.8426682151678768, |
|
"learning_rate": 1.8821469157260687e-05, |
|
"loss": 2.1086, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.7063020214030915, |
|
"grad_norm": 1.0388273479050054, |
|
"learning_rate": 1.8812430968292888e-05, |
|
"loss": 2.3994, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.7086801426872771, |
|
"grad_norm": 0.9544054737799986, |
|
"learning_rate": 1.8803360441735533e-05, |
|
"loss": 2.0918, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.7110582639714625, |
|
"grad_norm": 0.9333443425947632, |
|
"learning_rate": 1.8794257610873307e-05, |
|
"loss": 2.2388, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.713436385255648, |
|
"grad_norm": 0.9411120276164652, |
|
"learning_rate": 1.8785122509109425e-05, |
|
"loss": 2.1787, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.7158145065398336, |
|
"grad_norm": 0.899942759501507, |
|
"learning_rate": 1.877595516996554e-05, |
|
"loss": 2.2007, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.718192627824019, |
|
"grad_norm": 0.952238040187914, |
|
"learning_rate": 1.8766755627081586e-05, |
|
"loss": 2.1491, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.7205707491082045, |
|
"grad_norm": 1.0704167928673332, |
|
"learning_rate": 1.875752391421568e-05, |
|
"loss": 2.4306, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.72294887039239, |
|
"grad_norm": 1.0126343580756334, |
|
"learning_rate": 1.8748260065243985e-05, |
|
"loss": 2.3258, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.7253269916765755, |
|
"grad_norm": 0.9203276329051738, |
|
"learning_rate": 1.8738964114160586e-05, |
|
"loss": 2.0469, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.727705112960761, |
|
"grad_norm": 1.0298534605384366, |
|
"learning_rate": 1.8729636095077368e-05, |
|
"loss": 2.3807, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.7300832342449465, |
|
"grad_norm": 0.9102516225979571, |
|
"learning_rate": 1.8720276042223896e-05, |
|
"loss": 2.0705, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.7324613555291319, |
|
"grad_norm": 0.8450936687346169, |
|
"learning_rate": 1.8710883989947278e-05, |
|
"loss": 1.9468, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.7348394768133175, |
|
"grad_norm": 10.37449210680128, |
|
"learning_rate": 1.870145997271206e-05, |
|
"loss": 2.2267, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.737217598097503, |
|
"grad_norm": 0.9380258905092027, |
|
"learning_rate": 1.8692004025100054e-05, |
|
"loss": 2.2396, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.7395957193816884, |
|
"grad_norm": 0.9949249727532327, |
|
"learning_rate": 1.868251618181028e-05, |
|
"loss": 2.2737, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.7419738406658739, |
|
"grad_norm": 0.8650417909396523, |
|
"learning_rate": 1.8672996477658767e-05, |
|
"loss": 2.2057, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.7443519619500595, |
|
"grad_norm": 0.9963224791105781, |
|
"learning_rate": 1.866344494757848e-05, |
|
"loss": 2.3958, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.746730083234245, |
|
"grad_norm": 0.9782743745256545, |
|
"learning_rate": 1.8653861626619166e-05, |
|
"loss": 2.105, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.7491082045184304, |
|
"grad_norm": 0.9991591823799537, |
|
"learning_rate": 1.8644246549947226e-05, |
|
"loss": 2.3339, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.7491082045184304, |
|
"eval_loss": 2.453089952468872, |
|
"eval_runtime": 65.8189, |
|
"eval_samples_per_second": 9.708, |
|
"eval_steps_per_second": 1.215, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.751486325802616, |
|
"grad_norm": 0.8595275246052216, |
|
"learning_rate": 1.8634599752845594e-05, |
|
"loss": 2.2405, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.7538644470868014, |
|
"grad_norm": 0.9508091435584168, |
|
"learning_rate": 1.86249212707136e-05, |
|
"loss": 2.0259, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.7562425683709869, |
|
"grad_norm": 1.007143450486727, |
|
"learning_rate": 1.861521113906684e-05, |
|
"loss": 2.2581, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.7586206896551724, |
|
"grad_norm": 0.8694388672608973, |
|
"learning_rate": 1.8605469393537062e-05, |
|
"loss": 2.1028, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.760998810939358, |
|
"grad_norm": 1.1523113038633086, |
|
"learning_rate": 1.8595696069872013e-05, |
|
"loss": 2.5605, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.7633769322235434, |
|
"grad_norm": 1.0752431273595597, |
|
"learning_rate": 1.8585891203935315e-05, |
|
"loss": 2.179, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.7657550535077289, |
|
"grad_norm": 1.4087896659381576, |
|
"learning_rate": 1.8576054831706348e-05, |
|
"loss": 2.3727, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.7681331747919143, |
|
"grad_norm": 1.0235487680455349, |
|
"learning_rate": 1.856618698928009e-05, |
|
"loss": 2.0918, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.7705112960760999, |
|
"grad_norm": 0.9904332707712578, |
|
"learning_rate": 1.8556287712867006e-05, |
|
"loss": 2.2627, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.7728894173602854, |
|
"grad_norm": 0.9900875612198691, |
|
"learning_rate": 1.8546357038792918e-05, |
|
"loss": 2.2281, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.7752675386444708, |
|
"grad_norm": 0.8882946865776852, |
|
"learning_rate": 1.8536395003498857e-05, |
|
"loss": 2.2966, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.7776456599286563, |
|
"grad_norm": 0.9917259649244573, |
|
"learning_rate": 1.8526401643540924e-05, |
|
"loss": 2.2903, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.7800237812128419, |
|
"grad_norm": 1.0110335104954997, |
|
"learning_rate": 1.8516376995590185e-05, |
|
"loss": 2.2732, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.7824019024970273, |
|
"grad_norm": 1.0613748791083453, |
|
"learning_rate": 1.8506321096432516e-05, |
|
"loss": 2.1801, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.7847800237812128, |
|
"grad_norm": 0.8280074594268549, |
|
"learning_rate": 1.849623398296846e-05, |
|
"loss": 2.0413, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.7871581450653984, |
|
"grad_norm": 1.061771296332105, |
|
"learning_rate": 1.84861156922131e-05, |
|
"loss": 2.3187, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.7895362663495838, |
|
"grad_norm": 1.003700059201312, |
|
"learning_rate": 1.8475966261295947e-05, |
|
"loss": 2.2117, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.7919143876337693, |
|
"grad_norm": 1.0346094081558943, |
|
"learning_rate": 1.8465785727460763e-05, |
|
"loss": 2.35, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.7942925089179548, |
|
"grad_norm": 1.0628441168006169, |
|
"learning_rate": 1.845557412806545e-05, |
|
"loss": 2.2907, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.7966706302021404, |
|
"grad_norm": 0.9035239279839531, |
|
"learning_rate": 1.8445331500581905e-05, |
|
"loss": 2.3114, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.7990487514863258, |
|
"grad_norm": 0.8520631758698989, |
|
"learning_rate": 1.8435057882595885e-05, |
|
"loss": 2.236, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.8014268727705113, |
|
"grad_norm": 0.9909579691053704, |
|
"learning_rate": 1.8424753311806867e-05, |
|
"loss": 2.164, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.8038049940546967, |
|
"grad_norm": 1.3497517322085937, |
|
"learning_rate": 1.8414417826027907e-05, |
|
"loss": 2.1371, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.8061831153388823, |
|
"grad_norm": 0.9733174734597209, |
|
"learning_rate": 1.840405146318552e-05, |
|
"loss": 2.2605, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.8085612366230678, |
|
"grad_norm": 0.991877145495908, |
|
"learning_rate": 1.8393654261319504e-05, |
|
"loss": 2.2597, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.8109393579072532, |
|
"grad_norm": 0.8567918308119495, |
|
"learning_rate": 1.8383226258582833e-05, |
|
"loss": 2.0034, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.8133174791914387, |
|
"grad_norm": 1.065116900935523, |
|
"learning_rate": 1.837276749324151e-05, |
|
"loss": 2.3297, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.8156956004756243, |
|
"grad_norm": 0.945929516741479, |
|
"learning_rate": 1.8362278003674417e-05, |
|
"loss": 2.1654, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.8180737217598097, |
|
"grad_norm": 1.2321854339549771, |
|
"learning_rate": 1.8351757828373183e-05, |
|
"loss": 2.3218, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.8204518430439952, |
|
"grad_norm": 0.8769538008656999, |
|
"learning_rate": 1.8341207005942033e-05, |
|
"loss": 1.9733, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.8228299643281808, |
|
"grad_norm": 0.9563552488015744, |
|
"learning_rate": 1.8330625575097663e-05, |
|
"loss": 1.8932, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.8252080856123662, |
|
"grad_norm": 0.9405864126962645, |
|
"learning_rate": 1.8320013574669083e-05, |
|
"loss": 2.2498, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.8275862068965517, |
|
"grad_norm": 0.910096309278826, |
|
"learning_rate": 1.8309371043597472e-05, |
|
"loss": 2.0247, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.8299643281807372, |
|
"grad_norm": 0.8784633584470368, |
|
"learning_rate": 1.829869802093606e-05, |
|
"loss": 1.9894, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.8323424494649228, |
|
"grad_norm": 0.8541120126588495, |
|
"learning_rate": 1.8287994545849948e-05, |
|
"loss": 2.0511, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.8347205707491082, |
|
"grad_norm": 0.9979547733824415, |
|
"learning_rate": 1.8277260657615993e-05, |
|
"loss": 2.0893, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.8370986920332937, |
|
"grad_norm": 1.0836986822911394, |
|
"learning_rate": 1.826649639562266e-05, |
|
"loss": 2.3076, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.8394768133174791, |
|
"grad_norm": 1.035582585906488, |
|
"learning_rate": 1.825570179936986e-05, |
|
"loss": 2.2928, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.8418549346016647, |
|
"grad_norm": 0.8458996980228227, |
|
"learning_rate": 1.8244876908468826e-05, |
|
"loss": 2.0421, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.8442330558858502, |
|
"grad_norm": 0.9377520676275963, |
|
"learning_rate": 1.8234021762641946e-05, |
|
"loss": 2.2872, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.8466111771700356, |
|
"grad_norm": 0.9662487818099018, |
|
"learning_rate": 1.8223136401722648e-05, |
|
"loss": 2.2357, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.8489892984542212, |
|
"grad_norm": 0.9373837869014561, |
|
"learning_rate": 1.8212220865655224e-05, |
|
"loss": 2.3807, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.8513674197384067, |
|
"grad_norm": 0.8767072116092074, |
|
"learning_rate": 1.8201275194494695e-05, |
|
"loss": 2.0453, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.8537455410225921, |
|
"grad_norm": 0.9164578385360742, |
|
"learning_rate": 1.8190299428406667e-05, |
|
"loss": 2.1177, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.8561236623067776, |
|
"grad_norm": 2.207352437074081, |
|
"learning_rate": 1.8179293607667177e-05, |
|
"loss": 2.2001, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.8585017835909632, |
|
"grad_norm": 0.983155025111798, |
|
"learning_rate": 1.8168257772662556e-05, |
|
"loss": 2.3009, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.8608799048751486, |
|
"grad_norm": 0.9734880504382539, |
|
"learning_rate": 1.8157191963889265e-05, |
|
"loss": 2.3093, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.8632580261593341, |
|
"grad_norm": 0.8887247626382502, |
|
"learning_rate": 1.8146096221953767e-05, |
|
"loss": 2.1673, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.8656361474435196, |
|
"grad_norm": 0.9468700496948975, |
|
"learning_rate": 1.8134970587572345e-05, |
|
"loss": 2.2193, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.8680142687277052, |
|
"grad_norm": 0.9774148680996411, |
|
"learning_rate": 1.8123815101570996e-05, |
|
"loss": 2.2185, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.8703923900118906, |
|
"grad_norm": 1.0377594097114105, |
|
"learning_rate": 1.8112629804885248e-05, |
|
"loss": 2.1385, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.8727705112960761, |
|
"grad_norm": 0.9257473414011718, |
|
"learning_rate": 1.8101414738560018e-05, |
|
"loss": 2.309, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.8751486325802615, |
|
"grad_norm": 0.8751605326561603, |
|
"learning_rate": 1.8090169943749477e-05, |
|
"loss": 2.0981, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.8775267538644471, |
|
"grad_norm": 0.8318450854470809, |
|
"learning_rate": 1.8078895461716867e-05, |
|
"loss": 2.0499, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.8799048751486326, |
|
"grad_norm": 0.9229990617872996, |
|
"learning_rate": 1.8067591333834382e-05, |
|
"loss": 2.0931, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.882282996432818, |
|
"grad_norm": 0.8885903764744792, |
|
"learning_rate": 1.8056257601583004e-05, |
|
"loss": 2.1094, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.8846611177170036, |
|
"grad_norm": 0.8862767802235494, |
|
"learning_rate": 1.8044894306552338e-05, |
|
"loss": 2.1633, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.8870392390011891, |
|
"grad_norm": 0.8644263027922633, |
|
"learning_rate": 1.8033501490440478e-05, |
|
"loss": 2.1869, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.8894173602853745, |
|
"grad_norm": 0.8908757391357947, |
|
"learning_rate": 1.802207919505385e-05, |
|
"loss": 2.0867, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.89179548156956, |
|
"grad_norm": 0.9706439808339724, |
|
"learning_rate": 1.801062746230705e-05, |
|
"loss": 2.2817, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.8941736028537456, |
|
"grad_norm": 0.9665648406193532, |
|
"learning_rate": 1.79991463342227e-05, |
|
"loss": 2.109, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.896551724137931, |
|
"grad_norm": 0.8628669010410803, |
|
"learning_rate": 1.798763585293128e-05, |
|
"loss": 2.0049, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.8989298454221165, |
|
"grad_norm": 0.9905082003227874, |
|
"learning_rate": 1.7976096060671e-05, |
|
"loss": 2.3004, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.901307966706302, |
|
"grad_norm": 0.9857769539636477, |
|
"learning_rate": 1.7964526999787606e-05, |
|
"loss": 2.2076, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.9036860879904876, |
|
"grad_norm": 0.9874365687027726, |
|
"learning_rate": 1.7952928712734266e-05, |
|
"loss": 2.2918, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.906064209274673, |
|
"grad_norm": 0.8854059768194814, |
|
"learning_rate": 1.7941301242071384e-05, |
|
"loss": 2.1416, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.9084423305588585, |
|
"grad_norm": 0.8891010909161526, |
|
"learning_rate": 1.792964463046646e-05, |
|
"loss": 2.2335, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.9108204518430439, |
|
"grad_norm": 0.8836139495461457, |
|
"learning_rate": 1.7917958920693923e-05, |
|
"loss": 2.0156, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.9131985731272295, |
|
"grad_norm": 0.862309723766035, |
|
"learning_rate": 1.790624415563498e-05, |
|
"loss": 2.2668, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.915576694411415, |
|
"grad_norm": 0.8736380448815635, |
|
"learning_rate": 1.7894500378277463e-05, |
|
"loss": 2.1338, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.9179548156956004, |
|
"grad_norm": 0.8450572821559386, |
|
"learning_rate": 1.7882727631715655e-05, |
|
"loss": 2.0142, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.920332936979786, |
|
"grad_norm": 0.9862096866699867, |
|
"learning_rate": 1.7870925959150155e-05, |
|
"loss": 1.9915, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.9227110582639715, |
|
"grad_norm": 1.0092700839816597, |
|
"learning_rate": 1.7859095403887697e-05, |
|
"loss": 2.0294, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.925089179548157, |
|
"grad_norm": 1.1074789492464077, |
|
"learning_rate": 1.7847236009341007e-05, |
|
"loss": 2.2817, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.9274673008323424, |
|
"grad_norm": 1.1905201839698938, |
|
"learning_rate": 1.7835347819028642e-05, |
|
"loss": 2.1793, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.929845422116528, |
|
"grad_norm": 0.9783879172577963, |
|
"learning_rate": 1.7823430876574815e-05, |
|
"loss": 2.0936, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.9322235434007135, |
|
"grad_norm": 0.8747590280954227, |
|
"learning_rate": 1.7811485225709255e-05, |
|
"loss": 2.0459, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.9346016646848989, |
|
"grad_norm": 0.9379273246769285, |
|
"learning_rate": 1.7799510910267032e-05, |
|
"loss": 2.3024, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.9369797859690844, |
|
"grad_norm": 1.0154785871015584, |
|
"learning_rate": 1.778750797418841e-05, |
|
"loss": 2.2336, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.93935790725327, |
|
"grad_norm": 2.334020052582292, |
|
"learning_rate": 1.7775476461518668e-05, |
|
"loss": 2.2146, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.9417360285374554, |
|
"grad_norm": 1.028424946167679, |
|
"learning_rate": 1.7763416416407953e-05, |
|
"loss": 2.148, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.9441141498216409, |
|
"grad_norm": 0.9939249719466411, |
|
"learning_rate": 1.7751327883111117e-05, |
|
"loss": 2.3384, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.9464922711058263, |
|
"grad_norm": 0.9622526498121815, |
|
"learning_rate": 1.773921090598754e-05, |
|
"loss": 2.2333, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.9488703923900119, |
|
"grad_norm": 0.9627114271614428, |
|
"learning_rate": 1.7727065529500986e-05, |
|
"loss": 2.0078, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.9512485136741974, |
|
"grad_norm": 1.0739248368345071, |
|
"learning_rate": 1.7714891798219432e-05, |
|
"loss": 2.431, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.9536266349583828, |
|
"grad_norm": 0.9459448330629742, |
|
"learning_rate": 1.7702689756814898e-05, |
|
"loss": 2.195, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.9560047562425684, |
|
"grad_norm": 0.985098727152756, |
|
"learning_rate": 1.7690459450063297e-05, |
|
"loss": 2.378, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.9583828775267539, |
|
"grad_norm": 0.9004480308740594, |
|
"learning_rate": 1.7678200922844256e-05, |
|
"loss": 2.0375, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.9607609988109393, |
|
"grad_norm": 1.0816635115771127, |
|
"learning_rate": 1.7665914220140964e-05, |
|
"loss": 2.2932, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.9631391200951248, |
|
"grad_norm": 0.856817467594963, |
|
"learning_rate": 1.7653599387039993e-05, |
|
"loss": 2.1395, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.9655172413793104, |
|
"grad_norm": 0.9816466176158317, |
|
"learning_rate": 1.764125646873115e-05, |
|
"loss": 2.2032, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.9678953626634959, |
|
"grad_norm": 1.0159180902253635, |
|
"learning_rate": 1.7628885510507295e-05, |
|
"loss": 2.3814, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.9702734839476813, |
|
"grad_norm": 0.8765185622300491, |
|
"learning_rate": 1.7616486557764187e-05, |
|
"loss": 2.1377, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.9726516052318668, |
|
"grad_norm": 0.8581340453429611, |
|
"learning_rate": 1.7604059656000313e-05, |
|
"loss": 2.1687, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.9750297265160524, |
|
"grad_norm": 1.0106580156311586, |
|
"learning_rate": 1.7591604850816705e-05, |
|
"loss": 2.4389, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.9774078478002378, |
|
"grad_norm": 0.9102932420528898, |
|
"learning_rate": 1.757912218791681e-05, |
|
"loss": 2.1774, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.9797859690844233, |
|
"grad_norm": 0.8897815759878576, |
|
"learning_rate": 1.7566611713106287e-05, |
|
"loss": 2.0938, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.9821640903686087, |
|
"grad_norm": 0.8151179975871685, |
|
"learning_rate": 1.7554073472292854e-05, |
|
"loss": 2.0923, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.9845422116527943, |
|
"grad_norm": 0.9325421154931588, |
|
"learning_rate": 1.7541507511486114e-05, |
|
"loss": 2.2798, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.9869203329369798, |
|
"grad_norm": 0.9169184208068022, |
|
"learning_rate": 1.75289138767974e-05, |
|
"loss": 2.2266, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.9892984542211652, |
|
"grad_norm": 0.9539647094698775, |
|
"learning_rate": 1.7516292614439586e-05, |
|
"loss": 2.3459, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.9916765755053508, |
|
"grad_norm": 1.0372921455846975, |
|
"learning_rate": 1.7503643770726924e-05, |
|
"loss": 2.2152, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.9940546967895363, |
|
"grad_norm": 0.8903926236033542, |
|
"learning_rate": 1.7490967392074897e-05, |
|
"loss": 2.1959, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.9964328180737217, |
|
"grad_norm": 0.8491943369235393, |
|
"learning_rate": 1.7478263525000003e-05, |
|
"loss": 2.265, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.9988109393579072, |
|
"grad_norm": 0.9138223778798192, |
|
"learning_rate": 1.7465532216119628e-05, |
|
"loss": 2.2167, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.9988109393579072, |
|
"eval_loss": 2.4223339557647705, |
|
"eval_runtime": 65.3623, |
|
"eval_samples_per_second": 9.776, |
|
"eval_steps_per_second": 1.224, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.9138223778798192, |
|
"learning_rate": 1.7452773512151847e-05, |
|
"loss": 2.272, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 1.0023781212841856, |
|
"grad_norm": 1.628296924382809, |
|
"learning_rate": 1.7439987459915265e-05, |
|
"loss": 1.7689, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 1.004756242568371, |
|
"grad_norm": 1.0685556763602093, |
|
"learning_rate": 1.7427174106328846e-05, |
|
"loss": 1.952, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.0071343638525565, |
|
"grad_norm": 1.1222612546095956, |
|
"learning_rate": 1.7414333498411734e-05, |
|
"loss": 2.0925, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.009512485136742, |
|
"grad_norm": 1.0714061881839094, |
|
"learning_rate": 1.7401465683283085e-05, |
|
"loss": 1.9147, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.0118906064209274, |
|
"grad_norm": 1.7138310113119204, |
|
"learning_rate": 1.7388570708161895e-05, |
|
"loss": 2.0451, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.014268727705113, |
|
"grad_norm": 1.2402516993910184, |
|
"learning_rate": 1.7375648620366817e-05, |
|
"loss": 2.0889, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.0166468489892984, |
|
"grad_norm": 1.120370001804281, |
|
"learning_rate": 1.7362699467316005e-05, |
|
"loss": 2.0449, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.019024970273484, |
|
"grad_norm": 1.1942571073271016, |
|
"learning_rate": 1.7349723296526923e-05, |
|
"loss": 1.9913, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.0214030915576695, |
|
"grad_norm": 1.193074546908054, |
|
"learning_rate": 1.7336720155616186e-05, |
|
"loss": 2.0203, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.0237812128418549, |
|
"grad_norm": 0.8336179192625718, |
|
"learning_rate": 1.732369009229937e-05, |
|
"loss": 1.9839, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.0261593341260404, |
|
"grad_norm": 1.0440055245119049, |
|
"learning_rate": 1.731063315439084e-05, |
|
"loss": 1.9837, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.028537455410226, |
|
"grad_norm": 1.0017223079381312, |
|
"learning_rate": 1.729754938980359e-05, |
|
"loss": 2.0407, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.0309155766944114, |
|
"grad_norm": 1.0631773992935534, |
|
"learning_rate": 1.7284438846549045e-05, |
|
"loss": 2.0097, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.033293697978597, |
|
"grad_norm": 0.9980542007217309, |
|
"learning_rate": 1.7271301572736904e-05, |
|
"loss": 2.0821, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.0356718192627823, |
|
"grad_norm": 1.176402924330386, |
|
"learning_rate": 1.725813761657495e-05, |
|
"loss": 1.8594, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.0380499405469679, |
|
"grad_norm": 0.9752892126553369, |
|
"learning_rate": 1.7244947026368878e-05, |
|
"loss": 1.9673, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.0404280618311534, |
|
"grad_norm": 0.8689953910803522, |
|
"learning_rate": 1.723172985052212e-05, |
|
"loss": 1.9404, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.0428061831153388, |
|
"grad_norm": 1.2475813437261176, |
|
"learning_rate": 1.7218486137535663e-05, |
|
"loss": 2.0084, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.0451843043995244, |
|
"grad_norm": 1.0792429756941824, |
|
"learning_rate": 1.720521593600787e-05, |
|
"loss": 1.7713, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.04756242568371, |
|
"grad_norm": 0.9766534238863007, |
|
"learning_rate": 1.7191919294634318e-05, |
|
"loss": 1.9303, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.0499405469678953, |
|
"grad_norm": 1.0433496430768199, |
|
"learning_rate": 1.717859626220759e-05, |
|
"loss": 2.0188, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.0523186682520809, |
|
"grad_norm": 1.119386810190894, |
|
"learning_rate": 1.7165246887617124e-05, |
|
"loss": 1.9778, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.0546967895362664, |
|
"grad_norm": 0.9743730355346135, |
|
"learning_rate": 1.715187121984901e-05, |
|
"loss": 1.8941, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.0570749108204518, |
|
"grad_norm": 1.0796463722814744, |
|
"learning_rate": 1.7138469307985832e-05, |
|
"loss": 1.9573, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.0594530321046374, |
|
"grad_norm": 1.0491377417614562, |
|
"learning_rate": 1.7125041201206478e-05, |
|
"loss": 1.8715, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.0618311533888227, |
|
"grad_norm": 1.0722587649757946, |
|
"learning_rate": 1.711158694878595e-05, |
|
"loss": 1.892, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.0642092746730083, |
|
"grad_norm": 0.9251557954985203, |
|
"learning_rate": 1.7098106600095204e-05, |
|
"loss": 1.8922, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.0665873959571939, |
|
"grad_norm": 1.0363952883035685, |
|
"learning_rate": 1.708460020460095e-05, |
|
"loss": 1.9076, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 1.0689655172413792, |
|
"grad_norm": 1.0688439486649002, |
|
"learning_rate": 1.7071067811865477e-05, |
|
"loss": 1.9799, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.0713436385255648, |
|
"grad_norm": 1.1027957836331184, |
|
"learning_rate": 1.7057509471546476e-05, |
|
"loss": 1.9104, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 1.0737217598097504, |
|
"grad_norm": 1.618486595162837, |
|
"learning_rate": 1.7043925233396855e-05, |
|
"loss": 2.0171, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 1.0760998810939357, |
|
"grad_norm": 1.0017778693822061, |
|
"learning_rate": 1.703031514726455e-05, |
|
"loss": 2.0216, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.0784780023781213, |
|
"grad_norm": 1.0070735770943433, |
|
"learning_rate": 1.7016679263092352e-05, |
|
"loss": 2.0046, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 1.0808561236623069, |
|
"grad_norm": 1.1684678408376339, |
|
"learning_rate": 1.700301763091771e-05, |
|
"loss": 2.0023, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.0832342449464922, |
|
"grad_norm": 1.0525233453838523, |
|
"learning_rate": 1.6989330300872576e-05, |
|
"loss": 1.7791, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.0856123662306778, |
|
"grad_norm": 0.9255203454055275, |
|
"learning_rate": 1.6975617323183175e-05, |
|
"loss": 1.9353, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 1.0879904875148632, |
|
"grad_norm": 1.0726155742434451, |
|
"learning_rate": 1.6961878748169868e-05, |
|
"loss": 1.8358, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 1.0903686087990487, |
|
"grad_norm": 1.0856909362569107, |
|
"learning_rate": 1.694811462624694e-05, |
|
"loss": 2.0117, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 1.0927467300832343, |
|
"grad_norm": 1.0205381713143487, |
|
"learning_rate": 1.6934325007922418e-05, |
|
"loss": 1.9674, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.0951248513674197, |
|
"grad_norm": 1.2186466927066344, |
|
"learning_rate": 1.6920509943797893e-05, |
|
"loss": 2.0536, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 1.0975029726516052, |
|
"grad_norm": 1.0835683944583578, |
|
"learning_rate": 1.690666948456833e-05, |
|
"loss": 2.0184, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 1.0998810939357908, |
|
"grad_norm": 0.9540266090666372, |
|
"learning_rate": 1.689280368102188e-05, |
|
"loss": 1.7483, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 1.1022592152199762, |
|
"grad_norm": 1.1931606541151196, |
|
"learning_rate": 1.6878912584039698e-05, |
|
"loss": 1.8391, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 1.1046373365041617, |
|
"grad_norm": 0.9898663043482883, |
|
"learning_rate": 1.6864996244595757e-05, |
|
"loss": 2.0927, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.1070154577883473, |
|
"grad_norm": 1.1388354501174491, |
|
"learning_rate": 1.6851054713756653e-05, |
|
"loss": 2.1605, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 1.1093935790725327, |
|
"grad_norm": 1.0945133712632522, |
|
"learning_rate": 1.6837088042681423e-05, |
|
"loss": 1.9853, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 1.1117717003567182, |
|
"grad_norm": 1.1984401216196918, |
|
"learning_rate": 1.6823096282621366e-05, |
|
"loss": 2.0257, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.1141498216409036, |
|
"grad_norm": 0.8917099119377462, |
|
"learning_rate": 1.6809079484919834e-05, |
|
"loss": 1.8659, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 1.1165279429250892, |
|
"grad_norm": 1.0116433318578837, |
|
"learning_rate": 1.679503770101206e-05, |
|
"loss": 2.2598, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.1189060642092747, |
|
"grad_norm": 3.175396684459374, |
|
"learning_rate": 1.6780970982424966e-05, |
|
"loss": 1.9356, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 1.12128418549346, |
|
"grad_norm": 1.2346613722854758, |
|
"learning_rate": 1.6766879380776983e-05, |
|
"loss": 1.8155, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 1.1236623067776457, |
|
"grad_norm": 1.006693147566446, |
|
"learning_rate": 1.675276294777783e-05, |
|
"loss": 1.8518, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 1.1260404280618312, |
|
"grad_norm": 0.8894722137287147, |
|
"learning_rate": 1.6738621735228363e-05, |
|
"loss": 2.0505, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 1.1284185493460166, |
|
"grad_norm": 1.103577294552584, |
|
"learning_rate": 1.672445579502036e-05, |
|
"loss": 1.9912, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.1307966706302022, |
|
"grad_norm": 1.096718928499225, |
|
"learning_rate": 1.671026517913634e-05, |
|
"loss": 1.8938, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 1.1331747919143877, |
|
"grad_norm": 0.9817288455594879, |
|
"learning_rate": 1.669604993964937e-05, |
|
"loss": 1.8101, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 1.135552913198573, |
|
"grad_norm": 1.1460045828685286, |
|
"learning_rate": 1.668181012872288e-05, |
|
"loss": 2.2724, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 1.1379310344827587, |
|
"grad_norm": 1.1002427752644104, |
|
"learning_rate": 1.666754579861045e-05, |
|
"loss": 2.0842, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 1.140309155766944, |
|
"grad_norm": 1.0583766063889544, |
|
"learning_rate": 1.6653257001655652e-05, |
|
"loss": 2.006, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.1426872770511296, |
|
"grad_norm": 1.0862329799584287, |
|
"learning_rate": 1.6638943790291838e-05, |
|
"loss": 2.074, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 1.1450653983353152, |
|
"grad_norm": 1.215641711818309, |
|
"learning_rate": 1.6624606217041933e-05, |
|
"loss": 1.9628, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 1.1474435196195005, |
|
"grad_norm": 0.9889402798664048, |
|
"learning_rate": 1.661024433451828e-05, |
|
"loss": 2.0147, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.149821640903686, |
|
"grad_norm": 0.9856498299194657, |
|
"learning_rate": 1.6595858195422414e-05, |
|
"loss": 1.8529, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.1521997621878715, |
|
"grad_norm": 1.0678057323434258, |
|
"learning_rate": 1.6581447852544877e-05, |
|
"loss": 1.7709, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.154577883472057, |
|
"grad_norm": 0.8386080848257256, |
|
"learning_rate": 1.6567013358765045e-05, |
|
"loss": 1.9122, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.1569560047562426, |
|
"grad_norm": 1.0694865554443245, |
|
"learning_rate": 1.6552554767050898e-05, |
|
"loss": 1.8617, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.1593341260404282, |
|
"grad_norm": 1.9413752963807895, |
|
"learning_rate": 1.6538072130458853e-05, |
|
"loss": 1.9043, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.1617122473246135, |
|
"grad_norm": 1.04735712587821, |
|
"learning_rate": 1.6523565502133562e-05, |
|
"loss": 1.8987, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.164090368608799, |
|
"grad_norm": 0.9832837853313156, |
|
"learning_rate": 1.6509034935307716e-05, |
|
"loss": 2.1317, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.1664684898929845, |
|
"grad_norm": 1.1810675679989917, |
|
"learning_rate": 1.6494480483301836e-05, |
|
"loss": 1.9238, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 1.16884661117717, |
|
"grad_norm": 0.9165494247402323, |
|
"learning_rate": 1.6479902199524116e-05, |
|
"loss": 2.0648, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.1712247324613556, |
|
"grad_norm": 0.9636466082435201, |
|
"learning_rate": 1.6465300137470178e-05, |
|
"loss": 2.0287, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 1.173602853745541, |
|
"grad_norm": 0.9807575958399394, |
|
"learning_rate": 1.645067435072291e-05, |
|
"loss": 1.8887, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.1759809750297265, |
|
"grad_norm": 0.9586275603088916, |
|
"learning_rate": 1.6436024892952256e-05, |
|
"loss": 1.8363, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.178359096313912, |
|
"grad_norm": 0.9457623737642574, |
|
"learning_rate": 1.6421351817915025e-05, |
|
"loss": 2.146, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.1807372175980975, |
|
"grad_norm": 1.0017147042513856, |
|
"learning_rate": 1.6406655179454694e-05, |
|
"loss": 2.0542, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 1.183115338882283, |
|
"grad_norm": 1.9322307606799163, |
|
"learning_rate": 1.6391935031501193e-05, |
|
"loss": 1.9879, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.1854934601664684, |
|
"grad_norm": 1.084335177141969, |
|
"learning_rate": 1.6377191428070734e-05, |
|
"loss": 2.0715, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 1.187871581450654, |
|
"grad_norm": 0.9850991570942521, |
|
"learning_rate": 1.63624244232656e-05, |
|
"loss": 1.9499, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.1902497027348395, |
|
"grad_norm": 1.0171080592281492, |
|
"learning_rate": 1.6347634071273932e-05, |
|
"loss": 1.9646, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.192627824019025, |
|
"grad_norm": 1.0777884529528277, |
|
"learning_rate": 1.6332820426369567e-05, |
|
"loss": 2.0598, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 1.1950059453032105, |
|
"grad_norm": 1.020533123546286, |
|
"learning_rate": 1.631798354291179e-05, |
|
"loss": 1.8633, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 1.197384066587396, |
|
"grad_norm": 1.2067764605837605, |
|
"learning_rate": 1.6303123475345182e-05, |
|
"loss": 1.9119, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.1997621878715814, |
|
"grad_norm": 1.053791154347935, |
|
"learning_rate": 1.6288240278199393e-05, |
|
"loss": 2.1143, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.202140309155767, |
|
"grad_norm": 1.0193158748451816, |
|
"learning_rate": 1.6273334006088946e-05, |
|
"loss": 1.9422, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 1.2045184304399523, |
|
"grad_norm": 1.1993210309578368, |
|
"learning_rate": 1.6258404713713035e-05, |
|
"loss": 1.9505, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.206896551724138, |
|
"grad_norm": 0.9278344614652998, |
|
"learning_rate": 1.624345245585534e-05, |
|
"loss": 1.8948, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.2092746730083235, |
|
"grad_norm": 0.8732547686659468, |
|
"learning_rate": 1.62284772873838e-05, |
|
"loss": 1.8943, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 1.2116527942925088, |
|
"grad_norm": 1.1985862989626987, |
|
"learning_rate": 1.6213479263250433e-05, |
|
"loss": 1.8923, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.2140309155766944, |
|
"grad_norm": 0.9559917937082356, |
|
"learning_rate": 1.6198458438491124e-05, |
|
"loss": 1.9478, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 1.21640903686088, |
|
"grad_norm": 1.0369642986640548, |
|
"learning_rate": 1.6183414868225434e-05, |
|
"loss": 1.932, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.2187871581450653, |
|
"grad_norm": 0.8541777304922422, |
|
"learning_rate": 1.616834860765637e-05, |
|
"loss": 2.0197, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.221165279429251, |
|
"grad_norm": 1.1533492531760159, |
|
"learning_rate": 1.6153259712070225e-05, |
|
"loss": 2.0733, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 1.2235434007134365, |
|
"grad_norm": 1.1544430837733473, |
|
"learning_rate": 1.613814823683634e-05, |
|
"loss": 1.9616, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.2259215219976218, |
|
"grad_norm": 0.904111613570178, |
|
"learning_rate": 1.6123014237406912e-05, |
|
"loss": 2.1678, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.2282996432818074, |
|
"grad_norm": 1.101310926122992, |
|
"learning_rate": 1.6107857769316798e-05, |
|
"loss": 1.8998, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.2306777645659928, |
|
"grad_norm": 0.9065269772612409, |
|
"learning_rate": 1.6092678888183298e-05, |
|
"loss": 2.0495, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.2330558858501783, |
|
"grad_norm": 1.252782833936401, |
|
"learning_rate": 1.6077477649705963e-05, |
|
"loss": 2.1285, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.235434007134364, |
|
"grad_norm": 0.9769860406730624, |
|
"learning_rate": 1.6062254109666383e-05, |
|
"loss": 2.0681, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.2378121284185493, |
|
"grad_norm": 1.1110445238181776, |
|
"learning_rate": 1.604700832392798e-05, |
|
"loss": 1.8933, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 1.2401902497027348, |
|
"grad_norm": 1.044536874093627, |
|
"learning_rate": 1.603174034843582e-05, |
|
"loss": 1.932, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.2425683709869204, |
|
"grad_norm": 1.271022221507985, |
|
"learning_rate": 1.601645023921638e-05, |
|
"loss": 2.0284, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 1.2449464922711058, |
|
"grad_norm": 1.1152833273519336, |
|
"learning_rate": 1.600113805237737e-05, |
|
"loss": 1.9726, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.2473246135552913, |
|
"grad_norm": 1.2380069046576747, |
|
"learning_rate": 1.5985803844107502e-05, |
|
"loss": 1.8106, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.2473246135552913, |
|
"eval_loss": 2.4183852672576904, |
|
"eval_runtime": 65.6332, |
|
"eval_samples_per_second": 9.736, |
|
"eval_steps_per_second": 1.219, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.249702734839477, |
|
"grad_norm": 1.4005597471855011, |
|
"learning_rate": 1.5970447670676314e-05, |
|
"loss": 1.9685, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.2520808561236623, |
|
"grad_norm": 0.9054766632630273, |
|
"learning_rate": 1.5955069588433932e-05, |
|
"loss": 1.9919, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 1.2544589774078478, |
|
"grad_norm": 1.1186292190334863, |
|
"learning_rate": 1.5939669653810882e-05, |
|
"loss": 1.8359, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.2568370986920332, |
|
"grad_norm": 0.8661419871842052, |
|
"learning_rate": 1.5924247923317874e-05, |
|
"loss": 2.0698, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 1.2592152199762188, |
|
"grad_norm": 1.3186144918707836, |
|
"learning_rate": 1.5908804453545608e-05, |
|
"loss": 1.8845, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.2615933412604043, |
|
"grad_norm": 1.3508521915293898, |
|
"learning_rate": 1.589333930116455e-05, |
|
"loss": 2.0008, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.2639714625445897, |
|
"grad_norm": 1.0426903382069417, |
|
"learning_rate": 1.5877852522924733e-05, |
|
"loss": 2.14, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.2663495838287753, |
|
"grad_norm": 1.1270060025806774, |
|
"learning_rate": 1.5862344175655547e-05, |
|
"loss": 1.7585, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 1.2687277051129606, |
|
"grad_norm": 0.894696959523861, |
|
"learning_rate": 1.584681431626553e-05, |
|
"loss": 1.7954, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.2711058263971462, |
|
"grad_norm": 0.8963443148345133, |
|
"learning_rate": 1.5831263001742168e-05, |
|
"loss": 1.8992, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.2734839476813318, |
|
"grad_norm": 0.9751034618814204, |
|
"learning_rate": 1.581569028915166e-05, |
|
"loss": 2.0839, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.2758620689655173, |
|
"grad_norm": 1.2846795383089968, |
|
"learning_rate": 1.5800096235638735e-05, |
|
"loss": 2.0306, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.2782401902497027, |
|
"grad_norm": 1.1344265063076209, |
|
"learning_rate": 1.5784480898426444e-05, |
|
"loss": 2.0588, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 1.2806183115338883, |
|
"grad_norm": 2.3160794394882807, |
|
"learning_rate": 1.5768844334815923e-05, |
|
"loss": 2.2461, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 1.2829964328180736, |
|
"grad_norm": 1.2198804497684061, |
|
"learning_rate": 1.5753186602186207e-05, |
|
"loss": 2.0165, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.2853745541022592, |
|
"grad_norm": 1.0333157223752394, |
|
"learning_rate": 1.573750775799401e-05, |
|
"loss": 1.9158, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 1.2877526753864448, |
|
"grad_norm": 1.0933945363135282, |
|
"learning_rate": 1.572180785977352e-05, |
|
"loss": 2.0026, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 1.2901307966706301, |
|
"grad_norm": 1.0383257582502903, |
|
"learning_rate": 1.570608696513618e-05, |
|
"loss": 2.1656, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.2925089179548157, |
|
"grad_norm": 1.238965226787101, |
|
"learning_rate": 1.5690345131770474e-05, |
|
"loss": 1.8335, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.294887039239001, |
|
"grad_norm": 0.7431372863827379, |
|
"learning_rate": 1.5674582417441734e-05, |
|
"loss": 2.0428, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.2972651605231866, |
|
"grad_norm": 1.482428714196644, |
|
"learning_rate": 1.5658798879991905e-05, |
|
"loss": 2.0023, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.2996432818073722, |
|
"grad_norm": 1.2061227422514962, |
|
"learning_rate": 1.564299457733935e-05, |
|
"loss": 2.0852, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.3020214030915578, |
|
"grad_norm": 1.2166153788730312, |
|
"learning_rate": 1.5627169567478627e-05, |
|
"loss": 1.7886, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.3043995243757431, |
|
"grad_norm": 0.9980529341355449, |
|
"learning_rate": 1.561132390848029e-05, |
|
"loss": 1.8238, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.3067776456599287, |
|
"grad_norm": 1.0516168634342369, |
|
"learning_rate": 1.5595457658490643e-05, |
|
"loss": 1.9991, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.309155766944114, |
|
"grad_norm": 1.1353824503332424, |
|
"learning_rate": 1.5579570875731572e-05, |
|
"loss": 2.2081, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.3115338882282996, |
|
"grad_norm": 1.0157342209528704, |
|
"learning_rate": 1.55636636185003e-05, |
|
"loss": 1.9142, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.3139120095124852, |
|
"grad_norm": 1.3048380581725043, |
|
"learning_rate": 1.5547735945169188e-05, |
|
"loss": 1.9015, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.3162901307966706, |
|
"grad_norm": 1.0849291353506971, |
|
"learning_rate": 1.55317879141855e-05, |
|
"loss": 1.892, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.3186682520808561, |
|
"grad_norm": 1.103994377178752, |
|
"learning_rate": 1.5515819584071216e-05, |
|
"loss": 2.0736, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.3210463733650415, |
|
"grad_norm": 0.87333743581362, |
|
"learning_rate": 1.5499831013422804e-05, |
|
"loss": 1.6585, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.323424494649227, |
|
"grad_norm": 1.5096036417812366, |
|
"learning_rate": 1.5483822260911002e-05, |
|
"loss": 2.0984, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.3258026159334126, |
|
"grad_norm": 0.9402021272644058, |
|
"learning_rate": 1.5467793385280602e-05, |
|
"loss": 1.919, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.3281807372175982, |
|
"grad_norm": 1.010147060155673, |
|
"learning_rate": 1.5451744445350252e-05, |
|
"loss": 1.789, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.3305588585017836, |
|
"grad_norm": 1.018932157068251, |
|
"learning_rate": 1.5435675500012212e-05, |
|
"loss": 1.9567, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.3329369797859691, |
|
"grad_norm": 0.9622353435965788, |
|
"learning_rate": 1.5419586608232163e-05, |
|
"loss": 2.102, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.3353151010701545, |
|
"grad_norm": 1.046337850818545, |
|
"learning_rate": 1.540347782904897e-05, |
|
"loss": 1.8589, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.33769322235434, |
|
"grad_norm": 1.0106090373600465, |
|
"learning_rate": 1.5387349221574493e-05, |
|
"loss": 2.0144, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.3400713436385256, |
|
"grad_norm": 1.1400911418818147, |
|
"learning_rate": 1.5371200844993332e-05, |
|
"loss": 1.9842, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.342449464922711, |
|
"grad_norm": 0.8684778766824857, |
|
"learning_rate": 1.535503275856264e-05, |
|
"loss": 2.0641, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.3448275862068966, |
|
"grad_norm": 0.9355737090349231, |
|
"learning_rate": 1.5338845021611906e-05, |
|
"loss": 1.8084, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 1.347205707491082, |
|
"grad_norm": 1.0306221444200083, |
|
"learning_rate": 1.5322637693542716e-05, |
|
"loss": 2.0131, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 1.3495838287752675, |
|
"grad_norm": 1.0214723689545668, |
|
"learning_rate": 1.5306410833828534e-05, |
|
"loss": 1.969, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.351961950059453, |
|
"grad_norm": 0.9625636615112916, |
|
"learning_rate": 1.529016450201453e-05, |
|
"loss": 1.9543, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 1.3543400713436387, |
|
"grad_norm": 0.9724729717579162, |
|
"learning_rate": 1.5273898757717295e-05, |
|
"loss": 1.929, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.356718192627824, |
|
"grad_norm": 1.0109447015124593, |
|
"learning_rate": 1.5257613660624673e-05, |
|
"loss": 1.9279, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 1.3590963139120096, |
|
"grad_norm": 1.008542064536025, |
|
"learning_rate": 1.5241309270495524e-05, |
|
"loss": 2.0551, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.361474435196195, |
|
"grad_norm": 1.2854483484165498, |
|
"learning_rate": 1.5224985647159489e-05, |
|
"loss": 2.0914, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 1.3638525564803805, |
|
"grad_norm": 0.9055884973780003, |
|
"learning_rate": 1.5208642850516806e-05, |
|
"loss": 1.8825, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 1.366230677764566, |
|
"grad_norm": 1.0375062326116673, |
|
"learning_rate": 1.5192280940538058e-05, |
|
"loss": 1.8674, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.3686087990487514, |
|
"grad_norm": 0.896010533212188, |
|
"learning_rate": 1.5175899977263963e-05, |
|
"loss": 2.0912, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.370986920332937, |
|
"grad_norm": 1.1501881622163808, |
|
"learning_rate": 1.5159500020805173e-05, |
|
"loss": 1.89, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 1.3733650416171224, |
|
"grad_norm": 0.9783540504581939, |
|
"learning_rate": 1.5143081131342015e-05, |
|
"loss": 1.9331, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 1.375743162901308, |
|
"grad_norm": 1.170097731492715, |
|
"learning_rate": 1.5126643369124303e-05, |
|
"loss": 2.0069, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 1.3781212841854935, |
|
"grad_norm": 1.0072974136421486, |
|
"learning_rate": 1.5110186794471105e-05, |
|
"loss": 2.0293, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.380499405469679, |
|
"grad_norm": 1.1697150901196478, |
|
"learning_rate": 1.5093711467770517e-05, |
|
"loss": 1.9685, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 1.3828775267538644, |
|
"grad_norm": 1.2206930372756868, |
|
"learning_rate": 1.5077217449479455e-05, |
|
"loss": 1.933, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 1.38525564803805, |
|
"grad_norm": 0.970392976217639, |
|
"learning_rate": 1.5060704800123413e-05, |
|
"loss": 2.0112, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 1.3876337693222354, |
|
"grad_norm": 1.0661843012528456, |
|
"learning_rate": 1.5044173580296267e-05, |
|
"loss": 1.9158, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.390011890606421, |
|
"grad_norm": 0.8514185489462422, |
|
"learning_rate": 1.502762385066002e-05, |
|
"loss": 1.8758, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.3923900118906065, |
|
"grad_norm": 0.9748541634753175, |
|
"learning_rate": 1.5011055671944616e-05, |
|
"loss": 2.1468, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 1.3947681331747919, |
|
"grad_norm": 0.8818979397080599, |
|
"learning_rate": 1.4994469104947686e-05, |
|
"loss": 1.9446, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 1.3971462544589774, |
|
"grad_norm": 1.167689162690369, |
|
"learning_rate": 1.4977864210534341e-05, |
|
"loss": 1.706, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.3995243757431628, |
|
"grad_norm": 0.9558567418999584, |
|
"learning_rate": 1.496124104963695e-05, |
|
"loss": 2.1427, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 1.4019024970273484, |
|
"grad_norm": 0.991232122819226, |
|
"learning_rate": 1.4944599683254903e-05, |
|
"loss": 2.0682, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.404280618311534, |
|
"grad_norm": 1.0316875876904197, |
|
"learning_rate": 1.4927940172454405e-05, |
|
"loss": 2.1085, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 1.4066587395957193, |
|
"grad_norm": 1.0732162496383806, |
|
"learning_rate": 1.4911262578368233e-05, |
|
"loss": 1.8279, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.4090368608799049, |
|
"grad_norm": 0.9257277902304547, |
|
"learning_rate": 1.4894566962195532e-05, |
|
"loss": 2.0153, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 1.4114149821640904, |
|
"grad_norm": 1.0669817139188487, |
|
"learning_rate": 1.4877853385201569e-05, |
|
"loss": 1.9106, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 1.4137931034482758, |
|
"grad_norm": 0.8934624324805909, |
|
"learning_rate": 1.4861121908717529e-05, |
|
"loss": 1.9279, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.4161712247324614, |
|
"grad_norm": 1.0784062752165995, |
|
"learning_rate": 1.4844372594140271e-05, |
|
"loss": 1.8346, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.418549346016647, |
|
"grad_norm": 1.0964764861814489, |
|
"learning_rate": 1.4827605502932118e-05, |
|
"loss": 2.1296, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 1.4209274673008323, |
|
"grad_norm": 15.356619023384987, |
|
"learning_rate": 1.4810820696620625e-05, |
|
"loss": 2.1312, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 1.4233055885850179, |
|
"grad_norm": 1.1118633779533458, |
|
"learning_rate": 1.4794018236798349e-05, |
|
"loss": 1.8619, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 1.4256837098692032, |
|
"grad_norm": 1.153501655154434, |
|
"learning_rate": 1.477719818512263e-05, |
|
"loss": 2.077, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.4280618311533888, |
|
"grad_norm": 1.0076021147835075, |
|
"learning_rate": 1.4760360603315362e-05, |
|
"loss": 1.9109, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 1.4304399524375744, |
|
"grad_norm": 0.9674703722137163, |
|
"learning_rate": 1.4743505553162765e-05, |
|
"loss": 2.0096, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 1.4328180737217597, |
|
"grad_norm": 1.0412828169607384, |
|
"learning_rate": 1.4726633096515163e-05, |
|
"loss": 1.9162, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 1.4351961950059453, |
|
"grad_norm": 0.8719798048430073, |
|
"learning_rate": 1.4709743295286751e-05, |
|
"loss": 1.9538, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.4375743162901307, |
|
"grad_norm": 1.1442918189307298, |
|
"learning_rate": 1.4692836211455373e-05, |
|
"loss": 1.9587, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.4399524375743162, |
|
"grad_norm": 1.023556986586939, |
|
"learning_rate": 1.4675911907062289e-05, |
|
"loss": 1.8141, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 1.4423305588585018, |
|
"grad_norm": 0.8486875820991789, |
|
"learning_rate": 1.4658970444211953e-05, |
|
"loss": 1.9474, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 1.4447086801426874, |
|
"grad_norm": 0.906460047909353, |
|
"learning_rate": 1.464201188507178e-05, |
|
"loss": 2.2903, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 1.4470868014268727, |
|
"grad_norm": 1.1791852257245115, |
|
"learning_rate": 1.4625036291871926e-05, |
|
"loss": 2.0229, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 1.4494649227110583, |
|
"grad_norm": 1.051198365641329, |
|
"learning_rate": 1.460804372690505e-05, |
|
"loss": 1.8857, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.4518430439952437, |
|
"grad_norm": 0.8672989481026866, |
|
"learning_rate": 1.4591034252526093e-05, |
|
"loss": 1.9414, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 1.4542211652794292, |
|
"grad_norm": 0.9235870505947725, |
|
"learning_rate": 1.4574007931152037e-05, |
|
"loss": 2.0249, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 1.4565992865636148, |
|
"grad_norm": 1.0522141111911232, |
|
"learning_rate": 1.4556964825261696e-05, |
|
"loss": 2.2079, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 1.4589774078478002, |
|
"grad_norm": 1.006897151013655, |
|
"learning_rate": 1.4539904997395468e-05, |
|
"loss": 1.9247, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 1.4613555291319857, |
|
"grad_norm": 0.9494032663031484, |
|
"learning_rate": 1.4522828510155123e-05, |
|
"loss": 1.9099, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.463733650416171, |
|
"grad_norm": 1.0076075073121573, |
|
"learning_rate": 1.4505735426203545e-05, |
|
"loss": 1.8666, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 1.4661117717003567, |
|
"grad_norm": 1.1408823580186684, |
|
"learning_rate": 1.4488625808264536e-05, |
|
"loss": 1.9044, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 1.4684898929845422, |
|
"grad_norm": 0.9085119845888443, |
|
"learning_rate": 1.4471499719122565e-05, |
|
"loss": 1.8932, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 1.4708680142687278, |
|
"grad_norm": 1.148726123633719, |
|
"learning_rate": 1.4454357221622546e-05, |
|
"loss": 1.9092, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 1.4732461355529132, |
|
"grad_norm": 0.9671542619932931, |
|
"learning_rate": 1.4437198378669598e-05, |
|
"loss": 1.9493, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.4756242568370987, |
|
"grad_norm": 1.2173093104656718, |
|
"learning_rate": 1.4420023253228825e-05, |
|
"loss": 2.2601, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 1.478002378121284, |
|
"grad_norm": 1.1927753831747965, |
|
"learning_rate": 1.4402831908325082e-05, |
|
"loss": 1.8732, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 1.4803804994054697, |
|
"grad_norm": 1.0640642993825382, |
|
"learning_rate": 1.4385624407042729e-05, |
|
"loss": 2.0229, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 1.4827586206896552, |
|
"grad_norm": 1.0742996295504268, |
|
"learning_rate": 1.4368400812525434e-05, |
|
"loss": 1.8983, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 1.4851367419738406, |
|
"grad_norm": 0.9228290280482974, |
|
"learning_rate": 1.4351161187975904e-05, |
|
"loss": 2.1249, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.4875148632580262, |
|
"grad_norm": 0.9738223203214996, |
|
"learning_rate": 1.4333905596655668e-05, |
|
"loss": 1.9615, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 1.4898929845422115, |
|
"grad_norm": 1.0847722352198572, |
|
"learning_rate": 1.4316634101884859e-05, |
|
"loss": 2.0659, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 1.492271105826397, |
|
"grad_norm": 1.2154947037501092, |
|
"learning_rate": 1.4299346767041956e-05, |
|
"loss": 2.0163, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 1.4946492271105827, |
|
"grad_norm": 1.1111031435058485, |
|
"learning_rate": 1.4282043655563566e-05, |
|
"loss": 2.1607, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 1.4970273483947683, |
|
"grad_norm": 1.225646636267743, |
|
"learning_rate": 1.4264724830944198e-05, |
|
"loss": 2.0063, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.4970273483947683, |
|
"eval_loss": 2.4099409580230713, |
|
"eval_runtime": 65.5151, |
|
"eval_samples_per_second": 9.753, |
|
"eval_steps_per_second": 1.221, |
|
"step": 630 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 1680, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 210, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.908138113695744e+17, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|