{ "best_metric": null, "best_model_checkpoint": null, "epoch": 9.998542037907015, "eval_steps": 500, "global_step": 30860, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.10011339705167666, "grad_norm": 0.018291741609573364, "learning_rate": 9.90051847051199e-05, "loss": 2.1839, "step": 309 }, { "epoch": 0.2002267941033533, "grad_norm": 0.02208893373608589, "learning_rate": 9.800388852883993e-05, "loss": 2.1361, "step": 618 }, { "epoch": 0.30034019115502997, "grad_norm": 0.012775925919413567, "learning_rate": 9.700259235255995e-05, "loss": 2.1625, "step": 927 }, { "epoch": 0.4004535882067066, "grad_norm": 0.016916805878281593, "learning_rate": 9.600129617627997e-05, "loss": 2.1534, "step": 1236 }, { "epoch": 0.5005669852583833, "grad_norm": 0.006638580933213234, "learning_rate": 9.5e-05, "loss": 2.1464, "step": 1545 }, { "epoch": 0.6006803823100599, "grad_norm": 0.013361390680074692, "learning_rate": 9.399870382372004e-05, "loss": 2.1685, "step": 1854 }, { "epoch": 0.7007937793617366, "grad_norm": 0.005704471841454506, "learning_rate": 9.300064808814e-05, "loss": 2.1539, "step": 2163 }, { "epoch": 0.8009071764134132, "grad_norm": 0.02282548137009144, "learning_rate": 9.199935191186002e-05, "loss": 2.1799, "step": 2472 }, { "epoch": 0.9010205734650899, "grad_norm": 0.011914879083633423, "learning_rate": 9.099805573558005e-05, "loss": 2.1487, "step": 2781 }, { "epoch": 1.000971974728657, "grad_norm": 0.020065952092409134, "learning_rate": 8.999675955930007e-05, "loss": 2.1416, "step": 3090 }, { "epoch": 1.1010853717803337, "grad_norm": 0.020726900547742844, "learning_rate": 8.899546338302009e-05, "loss": 2.1422, "step": 3399 }, { "epoch": 1.2011987688320103, "grad_norm": 0.02327028475701809, "learning_rate": 8.799416720674012e-05, "loss": 2.149, "step": 3708 }, { "epoch": 1.301312165883687, "grad_norm": 0.03631984442472458, "learning_rate": 8.699611147116008e-05, "loss": 2.1605, "step": 4017 }, { "epoch": 1.4014255629353638, "grad_norm": 0.03152529150247574, "learning_rate": 8.59948152948801e-05, "loss": 2.1436, "step": 4326 }, { "epoch": 1.5015389599870403, "grad_norm": 0.05229083448648453, "learning_rate": 8.499351911860013e-05, "loss": 2.1794, "step": 4635 }, { "epoch": 1.601652357038717, "grad_norm": 0.027535825967788696, "learning_rate": 8.399222294232017e-05, "loss": 2.1641, "step": 4944 }, { "epoch": 1.7017657540903937, "grad_norm": 0.0046606422401964664, "learning_rate": 8.299092676604019e-05, "loss": 2.1242, "step": 5253 }, { "epoch": 1.8018791511420704, "grad_norm": 0.040044769644737244, "learning_rate": 8.198963058976021e-05, "loss": 2.14, "step": 5562 }, { "epoch": 1.901992548193747, "grad_norm": 0.01644105464220047, "learning_rate": 8.098833441348024e-05, "loss": 2.1488, "step": 5871 }, { "epoch": 2.001943949457314, "grad_norm": 0.01979956403374672, "learning_rate": 7.99902786779002e-05, "loss": 2.1586, "step": 6180 }, { "epoch": 2.1020573465089907, "grad_norm": 0.027056917548179626, "learning_rate": 7.898898250162022e-05, "loss": 2.1503, "step": 6489 }, { "epoch": 2.2021707435606674, "grad_norm": 0.01640058308839798, "learning_rate": 7.798768632534025e-05, "loss": 2.1418, "step": 6798 }, { "epoch": 2.302284140612344, "grad_norm": 0.01607314869761467, "learning_rate": 7.698639014906027e-05, "loss": 2.1514, "step": 7107 }, { "epoch": 2.4023975376640205, "grad_norm": 0.017274878919124603, "learning_rate": 7.59850939727803e-05, "loss": 2.1465, "step": 7416 }, { "epoch": 2.5025109347156973, "grad_norm": 0.024440627545118332, "learning_rate": 7.498379779650033e-05, "loss": 2.1211, "step": 7725 }, { "epoch": 2.602624331767374, "grad_norm": 0.005127054639160633, "learning_rate": 7.39857420609203e-05, "loss": 2.1423, "step": 8034 }, { "epoch": 2.702737728819051, "grad_norm": 0.05081469565629959, "learning_rate": 7.298444588464032e-05, "loss": 2.1705, "step": 8343 }, { "epoch": 2.8028511258707276, "grad_norm": 0.01722005568444729, "learning_rate": 7.198314970836034e-05, "loss": 2.1462, "step": 8652 }, { "epoch": 2.902964522922404, "grad_norm": 0.03728850930929184, "learning_rate": 7.098185353208037e-05, "loss": 2.1306, "step": 8961 }, { "epoch": 3.002915924185971, "grad_norm": 0.016364697366952896, "learning_rate": 6.998055735580039e-05, "loss": 2.1639, "step": 9270 }, { "epoch": 3.103029321237648, "grad_norm": 0.005412334576249123, "learning_rate": 6.897926117952041e-05, "loss": 2.1624, "step": 9579 }, { "epoch": 3.2031427182893246, "grad_norm": 0.0071863215416669846, "learning_rate": 6.797796500324045e-05, "loss": 2.1682, "step": 9888 }, { "epoch": 3.303256115341001, "grad_norm": 0.01707269623875618, "learning_rate": 6.69799092676604e-05, "loss": 2.129, "step": 10197 }, { "epoch": 3.4033695123926777, "grad_norm": 0.0162820965051651, "learning_rate": 6.597861309138042e-05, "loss": 2.1552, "step": 10506 }, { "epoch": 3.5034829094443545, "grad_norm": 0.005164678208529949, "learning_rate": 6.497731691510046e-05, "loss": 2.137, "step": 10815 }, { "epoch": 3.6035963064960312, "grad_norm": 0.01631810888648033, "learning_rate": 6.397602073882049e-05, "loss": 2.1578, "step": 11124 }, { "epoch": 3.703709703547708, "grad_norm": 0.004154821392148733, "learning_rate": 6.297472456254051e-05, "loss": 2.1602, "step": 11433 }, { "epoch": 3.8038231005993843, "grad_norm": 0.004850070457905531, "learning_rate": 6.197342838626053e-05, "loss": 2.1412, "step": 11742 }, { "epoch": 3.903936497651061, "grad_norm": 0.0028279961552470922, "learning_rate": 6.0975372650680494e-05, "loss": 2.1338, "step": 12051 }, { "epoch": 4.003887898914628, "grad_norm": 0.016007574275135994, "learning_rate": 5.997407647440052e-05, "loss": 2.1753, "step": 12360 }, { "epoch": 4.104001295966305, "grad_norm": 0.013178674504160881, "learning_rate": 5.897278029812054e-05, "loss": 2.1533, "step": 12669 }, { "epoch": 4.204114693017981, "grad_norm": 0.010946434922516346, "learning_rate": 5.797148412184057e-05, "loss": 2.1506, "step": 12978 }, { "epoch": 4.304228090069659, "grad_norm": 0.06032814085483551, "learning_rate": 5.6970187945560595e-05, "loss": 2.1408, "step": 13287 }, { "epoch": 4.404341487121335, "grad_norm": 0.036340948194265366, "learning_rate": 5.596889176928063e-05, "loss": 2.1317, "step": 13596 }, { "epoch": 4.505750850477888, "grad_norm": 0.022454094141721725, "learning_rate": 5.4967595593000656e-05, "loss": 2.1524, "step": 13905 }, { "epoch": 4.605864247529564, "grad_norm": 0.01212249230593443, "learning_rate": 5.396953985742061e-05, "loss": 2.1681, "step": 14214 }, { "epoch": 4.705977644581241, "grad_norm": 0.021747123450040817, "learning_rate": 5.296824368114064e-05, "loss": 2.1562, "step": 14523 }, { "epoch": 4.806091041632918, "grad_norm": 0.005955239292234182, "learning_rate": 5.196694750486067e-05, "loss": 2.1828, "step": 14832 }, { "epoch": 4.906204438684594, "grad_norm": 0.012121310457587242, "learning_rate": 5.096565132858069e-05, "loss": 2.1437, "step": 15141 }, { "epoch": 5.006479831524381, "grad_norm": 0.015455316752195358, "learning_rate": 4.996435515230072e-05, "loss": 2.1528, "step": 15450 }, { "epoch": 5.106593228576057, "grad_norm": 0.0007723022717982531, "learning_rate": 4.896305897602074e-05, "loss": 2.1723, "step": 15759 }, { "epoch": 5.206706625627733, "grad_norm": 0.004882230423390865, "learning_rate": 4.7965003240440704e-05, "loss": 2.1525, "step": 16068 }, { "epoch": 5.30682002267941, "grad_norm": 0.012511253356933594, "learning_rate": 4.696370706416073e-05, "loss": 2.1458, "step": 16377 }, { "epoch": 5.406933419731087, "grad_norm": 0.0262750256806612, "learning_rate": 4.596241088788075e-05, "loss": 2.0936, "step": 16686 }, { "epoch": 5.507046816782764, "grad_norm": 0.007834335789084435, "learning_rate": 4.4961114711600785e-05, "loss": 2.1322, "step": 16995 }, { "epoch": 5.60716021383444, "grad_norm": 0.020112166181206703, "learning_rate": 4.3959818535320806e-05, "loss": 2.1223, "step": 17304 }, { "epoch": 5.707273610886117, "grad_norm": 0.0009047465864568949, "learning_rate": 4.295852235904083e-05, "loss": 2.1425, "step": 17613 }, { "epoch": 5.807387007937794, "grad_norm": 0.006675088778138161, "learning_rate": 4.195722618276085e-05, "loss": 2.1433, "step": 17922 }, { "epoch": 5.90750040498947, "grad_norm": 0.022665197029709816, "learning_rate": 4.0959170447180816e-05, "loss": 2.1942, "step": 18231 }, { "epoch": 6.007451806253037, "grad_norm": 0.039628468453884125, "learning_rate": 3.995787427090085e-05, "loss": 2.1444, "step": 18540 }, { "epoch": 6.107565203304714, "grad_norm": 0.0011520631378516555, "learning_rate": 3.895657809462087e-05, "loss": 2.1524, "step": 18849 }, { "epoch": 6.207678600356391, "grad_norm": 0.03555034101009369, "learning_rate": 3.79552819183409e-05, "loss": 2.1203, "step": 19158 }, { "epoch": 6.307791997408067, "grad_norm": 0.013503223657608032, "learning_rate": 3.695398574206092e-05, "loss": 2.1352, "step": 19467 }, { "epoch": 6.407905394459744, "grad_norm": 0.026100030168890953, "learning_rate": 3.595268956578095e-05, "loss": 2.1655, "step": 19776 }, { "epoch": 6.508018791511421, "grad_norm": 0.0006254952168092132, "learning_rate": 3.4954633830200915e-05, "loss": 2.1352, "step": 20085 }, { "epoch": 6.608132188563097, "grad_norm": 0.01487251278012991, "learning_rate": 3.3953337653920935e-05, "loss": 2.1799, "step": 20394 }, { "epoch": 6.708245585614774, "grad_norm": 0.0053134192712605, "learning_rate": 3.295204147764096e-05, "loss": 2.1506, "step": 20703 }, { "epoch": 6.808358982666451, "grad_norm": 0.006517978850752115, "learning_rate": 3.195074530136098e-05, "loss": 2.1567, "step": 21012 }, { "epoch": 6.908472379718127, "grad_norm": 0.014856001362204552, "learning_rate": 3.0949449125081016e-05, "loss": 2.1593, "step": 21321 }, { "epoch": 7.008423780981694, "grad_norm": 0.0019718091934919357, "learning_rate": 2.994815294880104e-05, "loss": 2.1634, "step": 21630 }, { "epoch": 7.108537178033371, "grad_norm": 0.017417173832654953, "learning_rate": 2.8946856772521063e-05, "loss": 2.1433, "step": 21939 }, { "epoch": 7.208650575085048, "grad_norm": 0.034113720059394836, "learning_rate": 2.7948801036941023e-05, "loss": 2.1489, "step": 22248 }, { "epoch": 7.308763972136725, "grad_norm": 0.018758224323391914, "learning_rate": 2.694750486066105e-05, "loss": 2.13, "step": 22557 }, { "epoch": 7.408877369188401, "grad_norm": 0.04419185221195221, "learning_rate": 2.594620868438108e-05, "loss": 2.1641, "step": 22866 }, { "epoch": 7.508990766240077, "grad_norm": 0.001091700978577137, "learning_rate": 2.4944912508101104e-05, "loss": 2.1763, "step": 23175 }, { "epoch": 7.609104163291755, "grad_norm": 0.0011129506165161729, "learning_rate": 2.3943616331821128e-05, "loss": 2.125, "step": 23484 }, { "epoch": 7.709217560343431, "grad_norm": 0.007060033269226551, "learning_rate": 2.2942320155541155e-05, "loss": 2.1357, "step": 23793 }, { "epoch": 7.809330957395108, "grad_norm": 0.004454698413610458, "learning_rate": 2.1944264419961118e-05, "loss": 2.1502, "step": 24102 }, { "epoch": 7.9094443544467845, "grad_norm": 0.014023036696016788, "learning_rate": 2.0942968243681142e-05, "loss": 2.1397, "step": 24411 }, { "epoch": 8.00939575571035, "grad_norm": 0.0020535311195999384, "learning_rate": 1.994167206740117e-05, "loss": 2.1594, "step": 24720 }, { "epoch": 8.109509152762028, "grad_norm": 0.030191343277692795, "learning_rate": 1.8940375891121192e-05, "loss": 2.1379, "step": 25029 }, { "epoch": 8.209622549813705, "grad_norm": 0.02658534049987793, "learning_rate": 1.793907971484122e-05, "loss": 2.163, "step": 25338 }, { "epoch": 8.30973594686538, "grad_norm": 0.006291504483669996, "learning_rate": 1.6937783538561243e-05, "loss": 2.1445, "step": 25647 }, { "epoch": 8.409849343917058, "grad_norm": 0.013936794362962246, "learning_rate": 1.593648736228127e-05, "loss": 2.1571, "step": 25956 }, { "epoch": 8.509962740968735, "grad_norm": 0.031892433762550354, "learning_rate": 1.4938431626701233e-05, "loss": 2.1562, "step": 26265 }, { "epoch": 8.610076138020412, "grad_norm": 0.0028675836510956287, "learning_rate": 1.3937135450421257e-05, "loss": 2.1373, "step": 26574 }, { "epoch": 8.710189535072088, "grad_norm": 0.027382852509617805, "learning_rate": 1.2935839274141284e-05, "loss": 2.1162, "step": 26883 }, { "epoch": 8.810302932123765, "grad_norm": 0.006504488177597523, "learning_rate": 1.193454309786131e-05, "loss": 2.1281, "step": 27192 }, { "epoch": 8.910416329175442, "grad_norm": 0.006998216733336449, "learning_rate": 1.0933246921581337e-05, "loss": 2.1681, "step": 27501 }, { "epoch": 9.01036773043901, "grad_norm": 0.0019873257260769606, "learning_rate": 9.931950745301362e-06, "loss": 2.1397, "step": 27810 }, { "epoch": 9.110481127490685, "grad_norm": 0.04183882847428322, "learning_rate": 8.933895009721323e-06, "loss": 2.1425, "step": 28119 }, { "epoch": 9.210594524542362, "grad_norm": 0.008224571123719215, "learning_rate": 7.932598833441349e-06, "loss": 2.1567, "step": 28428 }, { "epoch": 9.31070792159404, "grad_norm": 0.009601627476513386, "learning_rate": 6.931302657161375e-06, "loss": 2.1334, "step": 28737 }, { "epoch": 9.410821318645715, "grad_norm": 0.028333676978945732, "learning_rate": 5.9300064808814e-06, "loss": 2.1732, "step": 29046 }, { "epoch": 9.510934715697392, "grad_norm": 0.01850961521267891, "learning_rate": 4.9287103046014265e-06, "loss": 2.1401, "step": 29355 }, { "epoch": 9.611048112749069, "grad_norm": 0.0022975043393671513, "learning_rate": 3.927414128321452e-06, "loss": 2.1523, "step": 29664 }, { "epoch": 9.711161509800744, "grad_norm": 0.033216096460819244, "learning_rate": 2.9261179520414777e-06, "loss": 2.1497, "step": 29973 }, { "epoch": 9.811274906852422, "grad_norm": 0.024143142625689507, "learning_rate": 1.928062216461439e-06, "loss": 2.1504, "step": 30282 }, { "epoch": 9.911388303904099, "grad_norm": 0.0085253044962883, "learning_rate": 9.267660401814646e-07, "loss": 2.177, "step": 30591 }, { "epoch": 9.998542037907015, "step": 30860, "total_flos": 1.461291118888168e+18, "train_loss": 1.202193304075487, "train_runtime": 14493.7066, "train_samples_per_second": 255.536, "train_steps_per_second": 2.129 } ], "logging_steps": 309, "max_steps": 30860, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 6800, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.461291118888168e+18, "train_batch_size": 30, "trial_name": null, "trial_params": null }