|
{ |
|
"best_metric": 0.25790610909461975, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-200", |
|
"epoch": 1.0027726432532347, |
|
"eval_steps": 100, |
|
"global_step": 271, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0036968576709796672, |
|
"grad_norm": 6.672558307647705, |
|
"learning_rate": 5e-06, |
|
"loss": 1.8315, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0036968576709796672, |
|
"eval_loss": 0.4404342472553253, |
|
"eval_runtime": 67.183, |
|
"eval_samples_per_second": 6.787, |
|
"eval_steps_per_second": 1.697, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0073937153419593345, |
|
"grad_norm": 6.42157506942749, |
|
"learning_rate": 1e-05, |
|
"loss": 1.6851, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.011090573012939002, |
|
"grad_norm": 5.945183753967285, |
|
"learning_rate": 1.5e-05, |
|
"loss": 1.5755, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.014787430683918669, |
|
"grad_norm": 4.450586318969727, |
|
"learning_rate": 2e-05, |
|
"loss": 1.3689, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.018484288354898338, |
|
"grad_norm": 3.1546850204467773, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.1569, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.022181146025878003, |
|
"grad_norm": 2.7391388416290283, |
|
"learning_rate": 3e-05, |
|
"loss": 1.4513, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.025878003696857672, |
|
"grad_norm": 2.3573246002197266, |
|
"learning_rate": 3.5e-05, |
|
"loss": 1.3315, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.029574861367837338, |
|
"grad_norm": 2.5408079624176025, |
|
"learning_rate": 4e-05, |
|
"loss": 1.3033, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.033271719038817, |
|
"grad_norm": 2.607274293899536, |
|
"learning_rate": 4.5e-05, |
|
"loss": 1.4326, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.036968576709796676, |
|
"grad_norm": 2.4289093017578125, |
|
"learning_rate": 5e-05, |
|
"loss": 1.3362, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.04066543438077634, |
|
"grad_norm": 2.145063877105713, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 1.3975, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.04436229205175601, |
|
"grad_norm": 1.970733404159546, |
|
"learning_rate": 6e-05, |
|
"loss": 1.1519, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.04805914972273567, |
|
"grad_norm": 1.9928206205368042, |
|
"learning_rate": 6.500000000000001e-05, |
|
"loss": 1.1993, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.051756007393715345, |
|
"grad_norm": 1.8372405767440796, |
|
"learning_rate": 7e-05, |
|
"loss": 1.1659, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.05545286506469501, |
|
"grad_norm": 1.7766002416610718, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 1.1062, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.059149722735674676, |
|
"grad_norm": 2.722572088241577, |
|
"learning_rate": 8e-05, |
|
"loss": 1.0653, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.06284658040665435, |
|
"grad_norm": 1.769493818283081, |
|
"learning_rate": 8.5e-05, |
|
"loss": 1.0896, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.066543438077634, |
|
"grad_norm": 1.829550862312317, |
|
"learning_rate": 9e-05, |
|
"loss": 1.0492, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.07024029574861368, |
|
"grad_norm": 1.9306679964065552, |
|
"learning_rate": 9.5e-05, |
|
"loss": 1.2088, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.07393715341959335, |
|
"grad_norm": 1.6777886152267456, |
|
"learning_rate": 0.0001, |
|
"loss": 1.0575, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.07763401109057301, |
|
"grad_norm": 1.731042742729187, |
|
"learning_rate": 9.999608360361113e-05, |
|
"loss": 1.1284, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.08133086876155268, |
|
"grad_norm": 1.768632173538208, |
|
"learning_rate": 9.998433502797095e-05, |
|
"loss": 1.2378, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.08502772643253234, |
|
"grad_norm": 1.6279608011245728, |
|
"learning_rate": 9.996475611356264e-05, |
|
"loss": 1.0409, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.08872458410351201, |
|
"grad_norm": 1.7023952007293701, |
|
"learning_rate": 9.993734992753777e-05, |
|
"loss": 1.0546, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.09242144177449169, |
|
"grad_norm": 1.719754934310913, |
|
"learning_rate": 9.990212076323586e-05, |
|
"loss": 1.0298, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.09611829944547134, |
|
"grad_norm": 1.7114758491516113, |
|
"learning_rate": 9.98590741395118e-05, |
|
"loss": 1.0207, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.09981515711645102, |
|
"grad_norm": 1.8310364484786987, |
|
"learning_rate": 9.980821679987125e-05, |
|
"loss": 1.2518, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.10351201478743069, |
|
"grad_norm": 1.6704847812652588, |
|
"learning_rate": 9.974955671141424e-05, |
|
"loss": 1.0702, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.10720887245841035, |
|
"grad_norm": 1.610925555229187, |
|
"learning_rate": 9.968310306358715e-05, |
|
"loss": 1.0744, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.11090573012939002, |
|
"grad_norm": 1.8064525127410889, |
|
"learning_rate": 9.960886626674302e-05, |
|
"loss": 1.1032, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.11460258780036968, |
|
"grad_norm": 1.691179871559143, |
|
"learning_rate": 9.952685795051077e-05, |
|
"loss": 0.9404, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.11829944547134935, |
|
"grad_norm": 1.6878719329833984, |
|
"learning_rate": 9.943709096197335e-05, |
|
"loss": 1.0445, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.12199630314232902, |
|
"grad_norm": 1.616434931755066, |
|
"learning_rate": 9.933957936365515e-05, |
|
"loss": 1.0305, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.1256931608133087, |
|
"grad_norm": 1.6706738471984863, |
|
"learning_rate": 9.923433843131901e-05, |
|
"loss": 1.0578, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.12939001848428835, |
|
"grad_norm": 1.7515959739685059, |
|
"learning_rate": 9.912138465157325e-05, |
|
"loss": 1.0298, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.133086876155268, |
|
"grad_norm": 1.8046168088912964, |
|
"learning_rate": 9.900073571928886e-05, |
|
"loss": 1.1877, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.1367837338262477, |
|
"grad_norm": 1.6414388418197632, |
|
"learning_rate": 9.887241053482757e-05, |
|
"loss": 0.9156, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.14048059149722736, |
|
"grad_norm": 1.7848786115646362, |
|
"learning_rate": 9.873642920108091e-05, |
|
"loss": 1.0424, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.14417744916820702, |
|
"grad_norm": 1.8037313222885132, |
|
"learning_rate": 9.859281302032106e-05, |
|
"loss": 0.9362, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.1478743068391867, |
|
"grad_norm": 1.642606258392334, |
|
"learning_rate": 9.844158449086371e-05, |
|
"loss": 0.9545, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.15157116451016636, |
|
"grad_norm": 1.6696778535842896, |
|
"learning_rate": 9.828276730354353e-05, |
|
"loss": 0.9483, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.15526802218114602, |
|
"grad_norm": 1.7426838874816895, |
|
"learning_rate": 9.811638633800287e-05, |
|
"loss": 0.9966, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.1589648798521257, |
|
"grad_norm": 1.6507072448730469, |
|
"learning_rate": 9.79424676587942e-05, |
|
"loss": 0.8627, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.16266173752310537, |
|
"grad_norm": 1.8130640983581543, |
|
"learning_rate": 9.776103851129706e-05, |
|
"loss": 1.0294, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.16635859519408502, |
|
"grad_norm": 1.9146381616592407, |
|
"learning_rate": 9.757212731744974e-05, |
|
"loss": 1.1166, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.17005545286506468, |
|
"grad_norm": 1.9425106048583984, |
|
"learning_rate": 9.737576367129694e-05, |
|
"loss": 1.031, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.17375231053604437, |
|
"grad_norm": 1.9324619770050049, |
|
"learning_rate": 9.717197833435367e-05, |
|
"loss": 0.9654, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.17744916820702403, |
|
"grad_norm": 1.8076626062393188, |
|
"learning_rate": 9.696080323078621e-05, |
|
"loss": 0.7715, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.18114602587800369, |
|
"grad_norm": 2.016756057739258, |
|
"learning_rate": 9.67422714424111e-05, |
|
"loss": 0.9425, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.18484288354898337, |
|
"grad_norm": 2.1553258895874023, |
|
"learning_rate": 9.65164172035126e-05, |
|
"loss": 1.0886, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.18853974121996303, |
|
"grad_norm": 1.7580000162124634, |
|
"learning_rate": 9.628327589547976e-05, |
|
"loss": 1.1776, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.1922365988909427, |
|
"grad_norm": 1.4649349451065063, |
|
"learning_rate": 9.604288404126363e-05, |
|
"loss": 1.1552, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.19593345656192238, |
|
"grad_norm": 1.3294196128845215, |
|
"learning_rate": 9.579527929965582e-05, |
|
"loss": 0.9731, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.19963031423290203, |
|
"grad_norm": 1.3747055530548096, |
|
"learning_rate": 9.554050045938893e-05, |
|
"loss": 1.2221, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.2033271719038817, |
|
"grad_norm": 1.3121620416641235, |
|
"learning_rate": 9.52785874330602e-05, |
|
"loss": 1.059, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.20702402957486138, |
|
"grad_norm": 1.3893884420394897, |
|
"learning_rate": 9.500958125087882e-05, |
|
"loss": 1.1146, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.21072088724584104, |
|
"grad_norm": 1.3223919868469238, |
|
"learning_rate": 9.473352405423844e-05, |
|
"loss": 1.0077, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.2144177449168207, |
|
"grad_norm": 1.3322027921676636, |
|
"learning_rate": 9.445045908911536e-05, |
|
"loss": 1.0703, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.21811460258780038, |
|
"grad_norm": 1.4537849426269531, |
|
"learning_rate": 9.416043069929388e-05, |
|
"loss": 1.2211, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.22181146025878004, |
|
"grad_norm": 1.342868447303772, |
|
"learning_rate": 9.386348431941952e-05, |
|
"loss": 1.0331, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.2255083179297597, |
|
"grad_norm": 1.5239349603652954, |
|
"learning_rate": 9.355966646788151e-05, |
|
"loss": 1.1679, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.22920517560073936, |
|
"grad_norm": 1.3578258752822876, |
|
"learning_rate": 9.324902473952528e-05, |
|
"loss": 0.9844, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.23290203327171904, |
|
"grad_norm": 1.3773984909057617, |
|
"learning_rate": 9.293160779819659e-05, |
|
"loss": 1.0466, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.2365988909426987, |
|
"grad_norm": 1.3751522302627563, |
|
"learning_rate": 9.26074653691179e-05, |
|
"loss": 1.119, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.24029574861367836, |
|
"grad_norm": 1.328293800354004, |
|
"learning_rate": 9.227664823109883e-05, |
|
"loss": 1.0232, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.24399260628465805, |
|
"grad_norm": 1.310364007949829, |
|
"learning_rate": 9.193920820858112e-05, |
|
"loss": 0.9279, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.2476894639556377, |
|
"grad_norm": 1.508972406387329, |
|
"learning_rate": 9.15951981635202e-05, |
|
"loss": 1.0623, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.2513863216266174, |
|
"grad_norm": 1.4712318181991577, |
|
"learning_rate": 9.1244671987104e-05, |
|
"loss": 1.0572, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.25508317929759705, |
|
"grad_norm": 1.4435091018676758, |
|
"learning_rate": 9.08876845913106e-05, |
|
"loss": 1.087, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.2587800369685767, |
|
"grad_norm": 1.3524105548858643, |
|
"learning_rate": 9.052429190030589e-05, |
|
"loss": 0.9821, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.26247689463955637, |
|
"grad_norm": 1.4055194854736328, |
|
"learning_rate": 9.015455084168278e-05, |
|
"loss": 1.1205, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.266173752310536, |
|
"grad_norm": 1.4329872131347656, |
|
"learning_rate": 8.977851933754316e-05, |
|
"loss": 1.1594, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.2698706099815157, |
|
"grad_norm": 1.494642972946167, |
|
"learning_rate": 8.939625629542402e-05, |
|
"loss": 1.1341, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.2735674676524954, |
|
"grad_norm": 1.3670973777770996, |
|
"learning_rate": 8.900782159906927e-05, |
|
"loss": 0.8418, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.27726432532347506, |
|
"grad_norm": 1.4169329404830933, |
|
"learning_rate": 8.861327609904859e-05, |
|
"loss": 0.9392, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.2809611829944547, |
|
"grad_norm": 1.3964389562606812, |
|
"learning_rate": 8.821268160322482e-05, |
|
"loss": 1.0313, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.2846580406654344, |
|
"grad_norm": 1.298718810081482, |
|
"learning_rate": 8.780610086707148e-05, |
|
"loss": 0.8463, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.28835489833641403, |
|
"grad_norm": 1.4148890972137451, |
|
"learning_rate": 8.739359758384162e-05, |
|
"loss": 0.931, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.2920517560073937, |
|
"grad_norm": 1.4295769929885864, |
|
"learning_rate": 8.697523637458997e-05, |
|
"loss": 0.9263, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.2957486136783734, |
|
"grad_norm": 1.2890899181365967, |
|
"learning_rate": 8.655108277804975e-05, |
|
"loss": 0.8196, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.29944547134935307, |
|
"grad_norm": 1.439947247505188, |
|
"learning_rate": 8.612120324036548e-05, |
|
"loss": 0.9164, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.3031423290203327, |
|
"grad_norm": 1.3879884481430054, |
|
"learning_rate": 8.568566510468391e-05, |
|
"loss": 0.8937, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.3068391866913124, |
|
"grad_norm": 1.5276659727096558, |
|
"learning_rate": 8.524453660060434e-05, |
|
"loss": 1.1198, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.31053604436229204, |
|
"grad_norm": 1.4229848384857178, |
|
"learning_rate": 8.479788683348995e-05, |
|
"loss": 0.8978, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.3142329020332717, |
|
"grad_norm": 1.614538550376892, |
|
"learning_rate": 8.434578577364218e-05, |
|
"loss": 1.22, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.3179297597042514, |
|
"grad_norm": 1.6977165937423706, |
|
"learning_rate": 8.388830424533934e-05, |
|
"loss": 1.0567, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.32162661737523107, |
|
"grad_norm": 1.510601282119751, |
|
"learning_rate": 8.342551391574165e-05, |
|
"loss": 0.9535, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.32532347504621073, |
|
"grad_norm": 1.423045039176941, |
|
"learning_rate": 8.295748728366413e-05, |
|
"loss": 0.9619, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.3290203327171904, |
|
"grad_norm": 1.5929685831069946, |
|
"learning_rate": 8.248429766821926e-05, |
|
"loss": 1.0364, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.33271719038817005, |
|
"grad_norm": 1.4277887344360352, |
|
"learning_rate": 8.200601919733106e-05, |
|
"loss": 0.7981, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.3364140480591497, |
|
"grad_norm": 1.663561224937439, |
|
"learning_rate": 8.15227267961226e-05, |
|
"loss": 1.019, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.34011090573012936, |
|
"grad_norm": 1.484633445739746, |
|
"learning_rate": 8.103449617517851e-05, |
|
"loss": 0.9798, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.3438077634011091, |
|
"grad_norm": 1.6405881643295288, |
|
"learning_rate": 8.054140381868436e-05, |
|
"loss": 1.0093, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.34750462107208874, |
|
"grad_norm": 1.6323596239089966, |
|
"learning_rate": 8.004352697244517e-05, |
|
"loss": 0.9985, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.3512014787430684, |
|
"grad_norm": 1.4849313497543335, |
|
"learning_rate": 7.954094363178422e-05, |
|
"loss": 0.7756, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.35489833641404805, |
|
"grad_norm": 1.7660150527954102, |
|
"learning_rate": 7.903373252932473e-05, |
|
"loss": 1.0436, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.3585951940850277, |
|
"grad_norm": 1.62890625, |
|
"learning_rate": 7.852197312265592e-05, |
|
"loss": 0.8442, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.36229205175600737, |
|
"grad_norm": 1.7414716482162476, |
|
"learning_rate": 7.800574558188547e-05, |
|
"loss": 0.9646, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.3659889094269871, |
|
"grad_norm": 1.6968168020248413, |
|
"learning_rate": 7.748513077708044e-05, |
|
"loss": 0.8914, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.36968576709796674, |
|
"grad_norm": 1.9534008502960205, |
|
"learning_rate": 7.696021026559849e-05, |
|
"loss": 0.9362, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.36968576709796674, |
|
"eval_loss": 0.27733513712882996, |
|
"eval_runtime": 68.0174, |
|
"eval_samples_per_second": 6.704, |
|
"eval_steps_per_second": 1.676, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3733826247689464, |
|
"grad_norm": 1.5832418203353882, |
|
"learning_rate": 7.643106627931147e-05, |
|
"loss": 1.1038, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.37707948243992606, |
|
"grad_norm": 1.5433467626571655, |
|
"learning_rate": 7.589778171172322e-05, |
|
"loss": 1.2129, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.3807763401109057, |
|
"grad_norm": 1.270788311958313, |
|
"learning_rate": 7.536044010498395e-05, |
|
"loss": 1.0397, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.3844731977818854, |
|
"grad_norm": 1.1377924680709839, |
|
"learning_rate": 7.48191256368028e-05, |
|
"loss": 0.9277, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.38817005545286504, |
|
"grad_norm": 1.2180787324905396, |
|
"learning_rate": 7.427392310726088e-05, |
|
"loss": 1.1243, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.39186691312384475, |
|
"grad_norm": 1.5344997644424438, |
|
"learning_rate": 7.372491792552693e-05, |
|
"loss": 1.1278, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.3955637707948244, |
|
"grad_norm": 1.2985904216766357, |
|
"learning_rate": 7.317219609647739e-05, |
|
"loss": 1.1904, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.39926062846580407, |
|
"grad_norm": 1.1818190813064575, |
|
"learning_rate": 7.261584420722329e-05, |
|
"loss": 1.027, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.4029574861367837, |
|
"grad_norm": 1.2055531740188599, |
|
"learning_rate": 7.20559494135458e-05, |
|
"loss": 1.374, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.4066543438077634, |
|
"grad_norm": 1.2185813188552856, |
|
"learning_rate": 7.149259942624287e-05, |
|
"loss": 1.0247, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.41035120147874304, |
|
"grad_norm": 1.2887550592422485, |
|
"learning_rate": 7.092588249738871e-05, |
|
"loss": 1.0531, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.41404805914972276, |
|
"grad_norm": 1.2477582693099976, |
|
"learning_rate": 7.03558874065087e-05, |
|
"loss": 1.1258, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.4177449168207024, |
|
"grad_norm": 1.2704106569290161, |
|
"learning_rate": 6.978270344667143e-05, |
|
"loss": 1.0419, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.4214417744916821, |
|
"grad_norm": 1.257286548614502, |
|
"learning_rate": 6.920642041050055e-05, |
|
"loss": 1.0382, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.42513863216266173, |
|
"grad_norm": 1.2782636880874634, |
|
"learning_rate": 6.862712857610813e-05, |
|
"loss": 0.833, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.4288354898336414, |
|
"grad_norm": 1.3618171215057373, |
|
"learning_rate": 6.804491869295206e-05, |
|
"loss": 1.0615, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.43253234750462105, |
|
"grad_norm": 1.2854083776474, |
|
"learning_rate": 6.745988196761975e-05, |
|
"loss": 1.0245, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.43622920517560076, |
|
"grad_norm": 1.2890700101852417, |
|
"learning_rate": 6.687211004953992e-05, |
|
"loss": 1.0267, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.4399260628465804, |
|
"grad_norm": 1.2263938188552856, |
|
"learning_rate": 6.628169501662526e-05, |
|
"loss": 0.9602, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.4436229205175601, |
|
"grad_norm": 1.3618099689483643, |
|
"learning_rate": 6.56887293608479e-05, |
|
"loss": 1.167, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.44731977818853974, |
|
"grad_norm": 1.3458164930343628, |
|
"learning_rate": 6.509330597374993e-05, |
|
"loss": 1.1031, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.4510166358595194, |
|
"grad_norm": 1.395233392715454, |
|
"learning_rate": 6.44955181318915e-05, |
|
"loss": 1.0788, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.45471349353049906, |
|
"grad_norm": 1.3283040523529053, |
|
"learning_rate": 6.38954594822384e-05, |
|
"loss": 1.025, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.4584103512014787, |
|
"grad_norm": 1.4176121950149536, |
|
"learning_rate": 6.329322402749181e-05, |
|
"loss": 1.0014, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.46210720887245843, |
|
"grad_norm": 1.350151777267456, |
|
"learning_rate": 6.268890611136211e-05, |
|
"loss": 0.9543, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.4658040665434381, |
|
"grad_norm": 1.4662792682647705, |
|
"learning_rate": 6.208260040378946e-05, |
|
"loss": 1.1333, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.46950092421441775, |
|
"grad_norm": 1.4703161716461182, |
|
"learning_rate": 6.147440188611324e-05, |
|
"loss": 1.0494, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.4731977818853974, |
|
"grad_norm": 1.3620432615280151, |
|
"learning_rate": 6.086440583619257e-05, |
|
"loss": 0.9709, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.47689463955637706, |
|
"grad_norm": 1.547058343887329, |
|
"learning_rate": 6.025270781348055e-05, |
|
"loss": 1.0412, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.4805914972273567, |
|
"grad_norm": 1.2925318479537964, |
|
"learning_rate": 5.9639403644054246e-05, |
|
"loss": 0.8893, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.48428835489833644, |
|
"grad_norm": 1.461341381072998, |
|
"learning_rate": 5.9024589405603035e-05, |
|
"loss": 1.1617, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.4879852125693161, |
|
"grad_norm": 1.3552955389022827, |
|
"learning_rate": 5.840836141237748e-05, |
|
"loss": 0.9199, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.49168207024029575, |
|
"grad_norm": 1.3925375938415527, |
|
"learning_rate": 5.779081620010104e-05, |
|
"loss": 0.9524, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.4953789279112754, |
|
"grad_norm": 1.4026466608047485, |
|
"learning_rate": 5.717205051084731e-05, |
|
"loss": 1.0116, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.49907578558225507, |
|
"grad_norm": 1.488108515739441, |
|
"learning_rate": 5.655216127788472e-05, |
|
"loss": 1.0833, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.5027726432532348, |
|
"grad_norm": 1.3740253448486328, |
|
"learning_rate": 5.5931245610491415e-05, |
|
"loss": 0.8804, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.5064695009242144, |
|
"grad_norm": 1.4298604726791382, |
|
"learning_rate": 5.5309400778742484e-05, |
|
"loss": 0.9091, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.5101663585951941, |
|
"grad_norm": 1.3947893381118774, |
|
"learning_rate": 5.4686724198272075e-05, |
|
"loss": 0.8541, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.5138632162661737, |
|
"grad_norm": 1.494073510169983, |
|
"learning_rate": 5.406331341501264e-05, |
|
"loss": 0.9325, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.5175600739371534, |
|
"grad_norm": 1.4765574932098389, |
|
"learning_rate": 5.3439266089913796e-05, |
|
"loss": 0.863, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.5212569316081331, |
|
"grad_norm": 1.4526939392089844, |
|
"learning_rate": 5.281467998364314e-05, |
|
"loss": 0.8982, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.5249537892791127, |
|
"grad_norm": 1.6596426963806152, |
|
"learning_rate": 5.218965294127155e-05, |
|
"loss": 1.0352, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.5286506469500925, |
|
"grad_norm": 1.5690279006958008, |
|
"learning_rate": 5.156428287694508e-05, |
|
"loss": 0.871, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.532347504621072, |
|
"grad_norm": 1.5764353275299072, |
|
"learning_rate": 5.093866775854618e-05, |
|
"loss": 0.8132, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.5360443622920518, |
|
"grad_norm": 1.6206929683685303, |
|
"learning_rate": 5.0312905592346496e-05, |
|
"loss": 0.8991, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.5397412199630314, |
|
"grad_norm": 1.5904308557510376, |
|
"learning_rate": 4.9687094407653516e-05, |
|
"loss": 0.9002, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.5434380776340111, |
|
"grad_norm": 1.5194886922836304, |
|
"learning_rate": 4.9061332241453835e-05, |
|
"loss": 0.9315, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.5471349353049908, |
|
"grad_norm": 1.5949432849884033, |
|
"learning_rate": 4.843571712305493e-05, |
|
"loss": 0.8933, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.5508317929759704, |
|
"grad_norm": 1.750752568244934, |
|
"learning_rate": 4.7810347058728454e-05, |
|
"loss": 0.9813, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.5545286506469501, |
|
"grad_norm": 1.9377819299697876, |
|
"learning_rate": 4.718532001635687e-05, |
|
"loss": 0.9949, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.5582255083179297, |
|
"grad_norm": 1.2233893871307373, |
|
"learning_rate": 4.6560733910086215e-05, |
|
"loss": 0.9573, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.5619223659889094, |
|
"grad_norm": 1.3002984523773193, |
|
"learning_rate": 4.593668658498738e-05, |
|
"loss": 1.0886, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.5656192236598891, |
|
"grad_norm": 1.234018087387085, |
|
"learning_rate": 4.531327580172794e-05, |
|
"loss": 0.9641, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.5693160813308688, |
|
"grad_norm": 1.2332717180252075, |
|
"learning_rate": 4.4690599221257534e-05, |
|
"loss": 1.0655, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.5730129390018485, |
|
"grad_norm": 1.1721649169921875, |
|
"learning_rate": 4.406875438950862e-05, |
|
"loss": 1.0076, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.5767097966728281, |
|
"grad_norm": 1.2083436250686646, |
|
"learning_rate": 4.34478387221153e-05, |
|
"loss": 1.1572, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.5804066543438078, |
|
"grad_norm": 1.1613600254058838, |
|
"learning_rate": 4.2827949489152716e-05, |
|
"loss": 1.0262, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.5841035120147874, |
|
"grad_norm": 1.1954678297042847, |
|
"learning_rate": 4.2209183799898975e-05, |
|
"loss": 1.0363, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.5878003696857671, |
|
"grad_norm": 1.1671994924545288, |
|
"learning_rate": 4.159163858762254e-05, |
|
"loss": 1.0286, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.5914972273567468, |
|
"grad_norm": 1.1261061429977417, |
|
"learning_rate": 4.097541059439698e-05, |
|
"loss": 0.9091, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5951940850277264, |
|
"grad_norm": 1.2204991579055786, |
|
"learning_rate": 4.036059635594578e-05, |
|
"loss": 1.1134, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.5988909426987061, |
|
"grad_norm": 1.129346489906311, |
|
"learning_rate": 3.9747292186519456e-05, |
|
"loss": 0.9158, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.6025878003696857, |
|
"grad_norm": 1.196499228477478, |
|
"learning_rate": 3.913559416380743e-05, |
|
"loss": 1.0781, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.6062846580406654, |
|
"grad_norm": 1.2142738103866577, |
|
"learning_rate": 3.8525598113886755e-05, |
|
"loss": 1.0086, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.609981515711645, |
|
"grad_norm": 1.1598280668258667, |
|
"learning_rate": 3.791739959621054e-05, |
|
"loss": 0.8981, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.6136783733826248, |
|
"grad_norm": 1.2825883626937866, |
|
"learning_rate": 3.73110938886379e-05, |
|
"loss": 1.0979, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.6173752310536045, |
|
"grad_norm": 1.2017385959625244, |
|
"learning_rate": 3.670677597250819e-05, |
|
"loss": 0.9349, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.6210720887245841, |
|
"grad_norm": 1.2944085597991943, |
|
"learning_rate": 3.610454051776159e-05, |
|
"loss": 1.028, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.6247689463955638, |
|
"grad_norm": 1.2570650577545166, |
|
"learning_rate": 3.5504481868108496e-05, |
|
"loss": 0.9435, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.6284658040665434, |
|
"grad_norm": 1.2900683879852295, |
|
"learning_rate": 3.490669402625007e-05, |
|
"loss": 1.0209, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.6321626617375231, |
|
"grad_norm": 1.3370497226715088, |
|
"learning_rate": 3.4311270639152125e-05, |
|
"loss": 1.0565, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.6358595194085028, |
|
"grad_norm": 1.3096612691879272, |
|
"learning_rate": 3.371830498337475e-05, |
|
"loss": 1.0703, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.6395563770794824, |
|
"grad_norm": 1.2441438436508179, |
|
"learning_rate": 3.31278899504601e-05, |
|
"loss": 0.9292, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.6432532347504621, |
|
"grad_norm": 1.3445340394973755, |
|
"learning_rate": 3.254011803238026e-05, |
|
"loss": 1.1239, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.6469500924214417, |
|
"grad_norm": 1.2935254573822021, |
|
"learning_rate": 3.195508130704795e-05, |
|
"loss": 0.9887, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.6506469500924215, |
|
"grad_norm": 1.3003994226455688, |
|
"learning_rate": 3.137287142389189e-05, |
|
"loss": 0.9951, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.6543438077634011, |
|
"grad_norm": 1.2665036916732788, |
|
"learning_rate": 3.079357958949946e-05, |
|
"loss": 0.9178, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.6580406654343808, |
|
"grad_norm": 1.3071720600128174, |
|
"learning_rate": 3.0217296553328578e-05, |
|
"loss": 0.9441, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.6617375231053605, |
|
"grad_norm": 1.3040660619735718, |
|
"learning_rate": 2.9644112593491313e-05, |
|
"loss": 0.9668, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.6654343807763401, |
|
"grad_norm": 1.3203717470169067, |
|
"learning_rate": 2.90741175026113e-05, |
|
"loss": 1.0138, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.6691312384473198, |
|
"grad_norm": 1.258781909942627, |
|
"learning_rate": 2.8507400573757158e-05, |
|
"loss": 0.8963, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.6728280961182994, |
|
"grad_norm": 1.2872486114501953, |
|
"learning_rate": 2.7944050586454214e-05, |
|
"loss": 0.8442, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.6765249537892791, |
|
"grad_norm": 1.318333387374878, |
|
"learning_rate": 2.738415579277672e-05, |
|
"loss": 0.9776, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.6802218114602587, |
|
"grad_norm": 1.4056696891784668, |
|
"learning_rate": 2.682780390352262e-05, |
|
"loss": 0.9503, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.6839186691312384, |
|
"grad_norm": 1.371435284614563, |
|
"learning_rate": 2.6275082074473077e-05, |
|
"loss": 0.8886, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.6876155268022182, |
|
"grad_norm": 1.3588190078735352, |
|
"learning_rate": 2.5726076892739125e-05, |
|
"loss": 0.8327, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.6913123844731978, |
|
"grad_norm": 1.4449481964111328, |
|
"learning_rate": 2.5180874363197215e-05, |
|
"loss": 1.0157, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.6950092421441775, |
|
"grad_norm": 1.4662877321243286, |
|
"learning_rate": 2.4639559895016068e-05, |
|
"loss": 0.9795, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.6987060998151571, |
|
"grad_norm": 1.4447919130325317, |
|
"learning_rate": 2.41022182882768e-05, |
|
"loss": 0.935, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.7024029574861368, |
|
"grad_norm": 1.4664056301116943, |
|
"learning_rate": 2.3568933720688545e-05, |
|
"loss": 0.9556, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.7060998151571165, |
|
"grad_norm": 1.4930390119552612, |
|
"learning_rate": 2.3039789734401522e-05, |
|
"loss": 0.9217, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.7097966728280961, |
|
"grad_norm": 1.5717326402664185, |
|
"learning_rate": 2.2514869222919572e-05, |
|
"loss": 1.0939, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.7134935304990758, |
|
"grad_norm": 1.5687856674194336, |
|
"learning_rate": 2.1994254418114522e-05, |
|
"loss": 1.0533, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.7171903881700554, |
|
"grad_norm": 1.4364888668060303, |
|
"learning_rate": 2.1478026877344087e-05, |
|
"loss": 0.8697, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.7208872458410351, |
|
"grad_norm": 1.6093854904174805, |
|
"learning_rate": 2.0966267470675273e-05, |
|
"loss": 1.062, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.7245841035120147, |
|
"grad_norm": 1.6295666694641113, |
|
"learning_rate": 2.0459056368215785e-05, |
|
"loss": 0.852, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.7282809611829945, |
|
"grad_norm": 1.6727817058563232, |
|
"learning_rate": 1.9956473027554846e-05, |
|
"loss": 0.9043, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.7319778188539742, |
|
"grad_norm": 1.708413004875183, |
|
"learning_rate": 1.945859618131564e-05, |
|
"loss": 0.8341, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.7356746765249538, |
|
"grad_norm": 1.630466341972351, |
|
"learning_rate": 1.8965503824821495e-05, |
|
"loss": 0.8544, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.7393715341959335, |
|
"grad_norm": 1.9450691938400269, |
|
"learning_rate": 1.8477273203877398e-05, |
|
"loss": 1.0403, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.7393715341959335, |
|
"eval_loss": 0.25790610909461975, |
|
"eval_runtime": 68.0375, |
|
"eval_samples_per_second": 6.702, |
|
"eval_steps_per_second": 1.676, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.7430683918669131, |
|
"grad_norm": 1.1571354866027832, |
|
"learning_rate": 1.7993980802668946e-05, |
|
"loss": 1.0016, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.7467652495378928, |
|
"grad_norm": 1.2866359949111938, |
|
"learning_rate": 1.7515702331780753e-05, |
|
"loss": 1.1153, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.7504621072088724, |
|
"grad_norm": 1.2495567798614502, |
|
"learning_rate": 1.7042512716335873e-05, |
|
"loss": 1.6495, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.7541589648798521, |
|
"grad_norm": 1.184030294418335, |
|
"learning_rate": 1.6574486084258366e-05, |
|
"loss": 0.9913, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.7578558225508318, |
|
"grad_norm": 1.1039363145828247, |
|
"learning_rate": 1.6111695754660667e-05, |
|
"loss": 0.9402, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.7615526802218114, |
|
"grad_norm": 1.1745617389678955, |
|
"learning_rate": 1.565421422635782e-05, |
|
"loss": 1.0854, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.7652495378927912, |
|
"grad_norm": 1.162819743156433, |
|
"learning_rate": 1.5202113166510057e-05, |
|
"loss": 1.1295, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.7689463955637708, |
|
"grad_norm": 1.0997729301452637, |
|
"learning_rate": 1.475546339939568e-05, |
|
"loss": 0.9278, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.7726432532347505, |
|
"grad_norm": 1.1672478914260864, |
|
"learning_rate": 1.4314334895316094e-05, |
|
"loss": 1.1381, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.7763401109057301, |
|
"grad_norm": 1.097519874572754, |
|
"learning_rate": 1.3878796759634544e-05, |
|
"loss": 0.9361, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.7800369685767098, |
|
"grad_norm": 1.0834153890609741, |
|
"learning_rate": 1.3448917221950263e-05, |
|
"loss": 0.8876, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.7837338262476895, |
|
"grad_norm": 1.1496264934539795, |
|
"learning_rate": 1.3024763625410024e-05, |
|
"loss": 0.9791, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.7874306839186691, |
|
"grad_norm": 1.1406208276748657, |
|
"learning_rate": 1.2606402416158392e-05, |
|
"loss": 0.964, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.7911275415896488, |
|
"grad_norm": 1.2033501863479614, |
|
"learning_rate": 1.2193899132928537e-05, |
|
"loss": 1.0953, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.7948243992606284, |
|
"grad_norm": 1.1828855276107788, |
|
"learning_rate": 1.1787318396775188e-05, |
|
"loss": 0.9958, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.7985212569316081, |
|
"grad_norm": 1.1799291372299194, |
|
"learning_rate": 1.138672390095143e-05, |
|
"loss": 1.0501, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.8022181146025879, |
|
"grad_norm": 1.232038974761963, |
|
"learning_rate": 1.0992178400930753e-05, |
|
"loss": 1.0139, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.8059149722735675, |
|
"grad_norm": 1.1742757558822632, |
|
"learning_rate": 1.060374370457599e-05, |
|
"loss": 0.8714, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.8096118299445472, |
|
"grad_norm": 1.2701990604400635, |
|
"learning_rate": 1.0221480662456845e-05, |
|
"loss": 0.9587, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.8133086876155268, |
|
"grad_norm": 1.295379400253296, |
|
"learning_rate": 9.845449158317215e-06, |
|
"loss": 1.1038, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.8170055452865065, |
|
"grad_norm": 1.2289679050445557, |
|
"learning_rate": 9.475708099694124e-06, |
|
"loss": 0.9174, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.8207024029574861, |
|
"grad_norm": 1.3198155164718628, |
|
"learning_rate": 9.112315408689414e-06, |
|
"loss": 1.0472, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.8243992606284658, |
|
"grad_norm": 1.1874481439590454, |
|
"learning_rate": 8.755328012896003e-06, |
|
"loss": 0.8252, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.8280961182994455, |
|
"grad_norm": 1.2251365184783936, |
|
"learning_rate": 8.404801836479808e-06, |
|
"loss": 0.8606, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.8317929759704251, |
|
"grad_norm": 1.3227380514144897, |
|
"learning_rate": 8.060791791418886e-06, |
|
"loss": 1.0696, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.8354898336414048, |
|
"grad_norm": 1.2983530759811401, |
|
"learning_rate": 7.723351768901172e-06, |
|
"loss": 0.9411, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.8391866913123844, |
|
"grad_norm": 1.256066083908081, |
|
"learning_rate": 7.392534630882092e-06, |
|
"loss": 0.8509, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.8428835489833642, |
|
"grad_norm": 1.2773510217666626, |
|
"learning_rate": 7.06839220180342e-06, |
|
"loss": 0.8596, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.8465804066543438, |
|
"grad_norm": 1.3085066080093384, |
|
"learning_rate": 6.750975260474718e-06, |
|
"loss": 0.885, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.8502772643253235, |
|
"grad_norm": 1.2915611267089844, |
|
"learning_rate": 6.440333532118503e-06, |
|
"loss": 0.9502, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.8539741219963032, |
|
"grad_norm": 1.2121251821517944, |
|
"learning_rate": 6.136515680580479e-06, |
|
"loss": 0.8245, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.8576709796672828, |
|
"grad_norm": 1.3207285404205322, |
|
"learning_rate": 5.839569300706127e-06, |
|
"loss": 0.8947, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.8613678373382625, |
|
"grad_norm": 1.4474835395812988, |
|
"learning_rate": 5.549540910884648e-06, |
|
"loss": 1.1095, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.8650646950092421, |
|
"grad_norm": 1.3800991773605347, |
|
"learning_rate": 5.266475945761562e-06, |
|
"loss": 0.8926, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.8687615526802218, |
|
"grad_norm": 1.2967448234558105, |
|
"learning_rate": 4.990418749121178e-06, |
|
"loss": 0.8284, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.8724584103512015, |
|
"grad_norm": 1.3374454975128174, |
|
"learning_rate": 4.721412566939804e-06, |
|
"loss": 0.9239, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.8761552680221811, |
|
"grad_norm": 1.404314637184143, |
|
"learning_rate": 4.459499540611078e-06, |
|
"loss": 0.9797, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.8798521256931608, |
|
"grad_norm": 1.361304759979248, |
|
"learning_rate": 4.2047207003442e-06, |
|
"loss": 0.8175, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.8835489833641405, |
|
"grad_norm": 1.464787244796753, |
|
"learning_rate": 3.9571159587363734e-06, |
|
"loss": 0.912, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.8872458410351202, |
|
"grad_norm": 1.3126565217971802, |
|
"learning_rate": 3.7167241045202473e-06, |
|
"loss": 0.8109, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.8909426987060998, |
|
"grad_norm": 1.3885515928268433, |
|
"learning_rate": 3.4835827964873945e-06, |
|
"loss": 0.8852, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.8946395563770795, |
|
"grad_norm": 1.48948335647583, |
|
"learning_rate": 3.2577285575889017e-06, |
|
"loss": 0.9546, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.8983364140480592, |
|
"grad_norm": 1.467980146408081, |
|
"learning_rate": 3.039196769213787e-06, |
|
"loss": 0.8802, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.9020332717190388, |
|
"grad_norm": 1.5345256328582764, |
|
"learning_rate": 2.8280216656463408e-06, |
|
"loss": 0.835, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.9057301293900185, |
|
"grad_norm": 1.5734572410583496, |
|
"learning_rate": 2.6242363287030613e-06, |
|
"loss": 1.0006, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.9094269870609981, |
|
"grad_norm": 1.4659297466278076, |
|
"learning_rate": 2.4278726825502697e-06, |
|
"loss": 0.8125, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.9131238447319778, |
|
"grad_norm": 1.6777136325836182, |
|
"learning_rate": 2.2389614887029566e-06, |
|
"loss": 0.9509, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.9168207024029574, |
|
"grad_norm": 1.7119040489196777, |
|
"learning_rate": 2.0575323412058033e-06, |
|
"loss": 0.8874, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.9205175600739371, |
|
"grad_norm": 1.8174391984939575, |
|
"learning_rate": 1.8836136619971466e-06, |
|
"loss": 0.9305, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.9242144177449169, |
|
"grad_norm": 1.8792368173599243, |
|
"learning_rate": 1.7172326964564778e-06, |
|
"loss": 0.9589, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.9279112754158965, |
|
"grad_norm": 1.1916697025299072, |
|
"learning_rate": 1.5584155091362906e-06, |
|
"loss": 1.0869, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.9316081330868762, |
|
"grad_norm": 1.109428882598877, |
|
"learning_rate": 1.4071869796789428e-06, |
|
"loss": 0.9649, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.9353049907578558, |
|
"grad_norm": 1.1600779294967651, |
|
"learning_rate": 1.263570798919106e-06, |
|
"loss": 1.0584, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.9390018484288355, |
|
"grad_norm": 1.027513027191162, |
|
"learning_rate": 1.1275894651724517e-06, |
|
"loss": 0.8654, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.9426987060998152, |
|
"grad_norm": 1.1321669816970825, |
|
"learning_rate": 9.992642807111485e-07, |
|
"loss": 0.9776, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.9463955637707948, |
|
"grad_norm": 1.2419346570968628, |
|
"learning_rate": 8.786153484267589e-07, |
|
"loss": 1.0648, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.9500924214417745, |
|
"grad_norm": 1.159534215927124, |
|
"learning_rate": 7.656615686809976e-07, |
|
"loss": 0.9293, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.9537892791127541, |
|
"grad_norm": 1.2147092819213867, |
|
"learning_rate": 6.604206363448661e-07, |
|
"loss": 0.9849, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.9574861367837338, |
|
"grad_norm": 1.2285771369934082, |
|
"learning_rate": 5.629090380266544e-07, |
|
"loss": 0.9584, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.9611829944547134, |
|
"grad_norm": 1.2933838367462158, |
|
"learning_rate": 4.7314204948923356e-07, |
|
"loss": 0.9436, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.9648798521256932, |
|
"grad_norm": 1.2568949460983276, |
|
"learning_rate": 3.9113373325698754e-07, |
|
"loss": 0.9507, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.9685767097966729, |
|
"grad_norm": 1.251380443572998, |
|
"learning_rate": 3.168969364128527e-07, |
|
"loss": 0.9798, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.9722735674676525, |
|
"grad_norm": 1.3915915489196777, |
|
"learning_rate": 2.5044328858576106e-07, |
|
"loss": 1.047, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.9759704251386322, |
|
"grad_norm": 1.353722333908081, |
|
"learning_rate": 1.917832001287645e-07, |
|
"loss": 0.8804, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.9796672828096118, |
|
"grad_norm": 1.3146427869796753, |
|
"learning_rate": 1.4092586048820578e-07, |
|
"loss": 0.9416, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.9833641404805915, |
|
"grad_norm": 1.4524424076080322, |
|
"learning_rate": 9.787923676414235e-08, |
|
"loss": 1.0211, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.9870609981515711, |
|
"grad_norm": 1.499550461769104, |
|
"learning_rate": 6.265007246223364e-08, |
|
"loss": 0.9393, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.9907578558225508, |
|
"grad_norm": 1.4518331289291382, |
|
"learning_rate": 3.524388643736387e-08, |
|
"loss": 0.8188, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.9944547134935305, |
|
"grad_norm": 1.5743730068206787, |
|
"learning_rate": 1.566497202904471e-08, |
|
"loss": 0.8282, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.9981515711645101, |
|
"grad_norm": 1.7006173133850098, |
|
"learning_rate": 3.91639638886998e-09, |
|
"loss": 0.7826, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.0027726432532347, |
|
"grad_norm": 1.2868958711624146, |
|
"learning_rate": 0.0, |
|
"loss": 0.9401, |
|
"step": 271 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 271, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 7.756009208713052e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|