|
{ |
|
"best_global_step": null, |
|
"best_metric": 0.8121369481086731, |
|
"best_model_checkpoint": null, |
|
"epoch": 5.915813424345847, |
|
"eval_steps": 200, |
|
"global_step": 10400, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02844141069397042, |
|
"grad_norm": 8.889737129211426, |
|
"learning_rate": 9.099526066350711e-07, |
|
"loss": 4.4241, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05688282138794084, |
|
"grad_norm": 7.543558120727539, |
|
"learning_rate": 1.8578199052132703e-06, |
|
"loss": 3.4415, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08532423208191127, |
|
"grad_norm": 7.774235725402832, |
|
"learning_rate": 2.8056872037914696e-06, |
|
"loss": 2.6725, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.11376564277588168, |
|
"grad_norm": 7.825632572174072, |
|
"learning_rate": 3.7535545023696683e-06, |
|
"loss": 2.4137, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.11376564277588168, |
|
"eval_loss": 2.2685751914978027, |
|
"eval_runtime": 29.7449, |
|
"eval_samples_per_second": 840.481, |
|
"eval_steps_per_second": 6.589, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.1422070534698521, |
|
"grad_norm": 8.4616060256958, |
|
"learning_rate": 4.701421800947868e-06, |
|
"loss": 2.2701, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.17064846416382254, |
|
"grad_norm": 7.439651966094971, |
|
"learning_rate": 5.6492890995260666e-06, |
|
"loss": 2.1523, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.19908987485779295, |
|
"grad_norm": 8.319734573364258, |
|
"learning_rate": 6.597156398104266e-06, |
|
"loss": 2.0805, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.22753128555176336, |
|
"grad_norm": 7.824019432067871, |
|
"learning_rate": 7.545023696682466e-06, |
|
"loss": 2.0513, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.22753128555176336, |
|
"eval_loss": 1.9506336450576782, |
|
"eval_runtime": 28.6984, |
|
"eval_samples_per_second": 871.127, |
|
"eval_steps_per_second": 6.83, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.25597269624573377, |
|
"grad_norm": 8.402134895324707, |
|
"learning_rate": 8.492890995260664e-06, |
|
"loss": 2.0048, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.2844141069397042, |
|
"grad_norm": 7.345431327819824, |
|
"learning_rate": 9.440758293838863e-06, |
|
"loss": 1.9552, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.31285551763367464, |
|
"grad_norm": 8.147149085998535, |
|
"learning_rate": 1.0388625592417063e-05, |
|
"loss": 1.8778, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.3412969283276451, |
|
"grad_norm": 7.802554130554199, |
|
"learning_rate": 1.133649289099526e-05, |
|
"loss": 1.8549, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.3412969283276451, |
|
"eval_loss": 1.7629565000534058, |
|
"eval_runtime": 33.6232, |
|
"eval_samples_per_second": 743.534, |
|
"eval_steps_per_second": 5.829, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.36973833902161546, |
|
"grad_norm": 7.983552932739258, |
|
"learning_rate": 1.228436018957346e-05, |
|
"loss": 1.822, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.3981797497155859, |
|
"grad_norm": 8.035250663757324, |
|
"learning_rate": 1.323222748815166e-05, |
|
"loss": 1.8128, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.42662116040955633, |
|
"grad_norm": 8.409351348876953, |
|
"learning_rate": 1.4180094786729858e-05, |
|
"loss": 1.7742, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.4550625711035267, |
|
"grad_norm": 7.7319183349609375, |
|
"learning_rate": 1.5127962085308059e-05, |
|
"loss": 1.7076, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.4550625711035267, |
|
"eval_loss": 1.6330854892730713, |
|
"eval_runtime": 33.0226, |
|
"eval_samples_per_second": 757.058, |
|
"eval_steps_per_second": 5.935, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.48350398179749715, |
|
"grad_norm": 7.466287136077881, |
|
"learning_rate": 1.6075829383886257e-05, |
|
"loss": 1.6919, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.5119453924914675, |
|
"grad_norm": 7.655446529388428, |
|
"learning_rate": 1.7023696682464458e-05, |
|
"loss": 1.64, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.540386803185438, |
|
"grad_norm": 8.173416137695312, |
|
"learning_rate": 1.7971563981042655e-05, |
|
"loss": 1.6291, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.5688282138794084, |
|
"grad_norm": 7.376980781555176, |
|
"learning_rate": 1.8919431279620855e-05, |
|
"loss": 1.5881, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.5688282138794084, |
|
"eval_loss": 1.5367897748947144, |
|
"eval_runtime": 32.9799, |
|
"eval_samples_per_second": 758.038, |
|
"eval_steps_per_second": 5.943, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.5972696245733788, |
|
"grad_norm": 7.863293170928955, |
|
"learning_rate": 1.9867298578199055e-05, |
|
"loss": 1.6018, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.6257110352673493, |
|
"grad_norm": 7.6200385093688965, |
|
"learning_rate": 1.9909406931423158e-05, |
|
"loss": 1.5664, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.6541524459613197, |
|
"grad_norm": 8.286286354064941, |
|
"learning_rate": 1.9804066154008218e-05, |
|
"loss": 1.5545, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.6825938566552902, |
|
"grad_norm": 7.845026969909668, |
|
"learning_rate": 1.969872537659328e-05, |
|
"loss": 1.5292, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.6825938566552902, |
|
"eval_loss": 1.4531670808792114, |
|
"eval_runtime": 29.4807, |
|
"eval_samples_per_second": 848.011, |
|
"eval_steps_per_second": 6.648, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.7110352673492605, |
|
"grad_norm": 7.120193004608154, |
|
"learning_rate": 1.9593384599178345e-05, |
|
"loss": 1.5166, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.7394766780432309, |
|
"grad_norm": 7.721842288970947, |
|
"learning_rate": 1.9488043821763408e-05, |
|
"loss": 1.517, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.7679180887372014, |
|
"grad_norm": 7.104468822479248, |
|
"learning_rate": 1.938270304434847e-05, |
|
"loss": 1.4639, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.7963594994311718, |
|
"grad_norm": 7.570240020751953, |
|
"learning_rate": 1.927736226693353e-05, |
|
"loss": 1.4729, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.7963594994311718, |
|
"eval_loss": 1.368685245513916, |
|
"eval_runtime": 28.6992, |
|
"eval_samples_per_second": 871.103, |
|
"eval_steps_per_second": 6.829, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.8248009101251422, |
|
"grad_norm": 7.745856761932373, |
|
"learning_rate": 1.9172021489518595e-05, |
|
"loss": 1.4501, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.8532423208191127, |
|
"grad_norm": 7.175948619842529, |
|
"learning_rate": 1.906668071210366e-05, |
|
"loss": 1.3932, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.8816837315130831, |
|
"grad_norm": 8.291092872619629, |
|
"learning_rate": 1.8961339934688722e-05, |
|
"loss": 1.4063, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.9101251422070534, |
|
"grad_norm": 7.994405269622803, |
|
"learning_rate": 1.8855999157273782e-05, |
|
"loss": 1.3825, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.9101251422070534, |
|
"eval_loss": 1.300325632095337, |
|
"eval_runtime": 28.6638, |
|
"eval_samples_per_second": 872.179, |
|
"eval_steps_per_second": 6.838, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.9385665529010239, |
|
"grad_norm": 8.009012222290039, |
|
"learning_rate": 1.8750658379858845e-05, |
|
"loss": 1.3647, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.9670079635949943, |
|
"grad_norm": 8.436450004577637, |
|
"learning_rate": 1.864531760244391e-05, |
|
"loss": 1.3431, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.9954493742889647, |
|
"grad_norm": 7.547204971313477, |
|
"learning_rate": 1.8539976825028972e-05, |
|
"loss": 1.3417, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.023890784982935, |
|
"grad_norm": 6.637471675872803, |
|
"learning_rate": 1.8434636047614032e-05, |
|
"loss": 1.0839, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.023890784982935, |
|
"eval_loss": 1.2430765628814697, |
|
"eval_runtime": 28.6828, |
|
"eval_samples_per_second": 871.603, |
|
"eval_steps_per_second": 6.833, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.0523321956769056, |
|
"grad_norm": 7.198896408081055, |
|
"learning_rate": 1.8329295270199096e-05, |
|
"loss": 1.0801, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.080773606370876, |
|
"grad_norm": 7.391284942626953, |
|
"learning_rate": 1.8223954492784156e-05, |
|
"loss": 1.0577, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.1092150170648465, |
|
"grad_norm": 6.571183681488037, |
|
"learning_rate": 1.811861371536922e-05, |
|
"loss": 1.0159, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.1376564277588168, |
|
"grad_norm": 7.20968770980835, |
|
"learning_rate": 1.8013272937954283e-05, |
|
"loss": 1.0239, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.1376564277588168, |
|
"eval_loss": 1.213191270828247, |
|
"eval_runtime": 28.5325, |
|
"eval_samples_per_second": 876.195, |
|
"eval_steps_per_second": 6.869, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.1660978384527874, |
|
"grad_norm": 6.97741174697876, |
|
"learning_rate": 1.7907932160539346e-05, |
|
"loss": 1.0335, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.1945392491467577, |
|
"grad_norm": 7.157691478729248, |
|
"learning_rate": 1.7802591383124406e-05, |
|
"loss": 1.0117, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.222980659840728, |
|
"grad_norm": 7.168184280395508, |
|
"learning_rate": 1.769725060570947e-05, |
|
"loss": 1.0343, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.2514220705346986, |
|
"grad_norm": 7.099086284637451, |
|
"learning_rate": 1.7591909828294533e-05, |
|
"loss": 1.0193, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.2514220705346986, |
|
"eval_loss": 1.1807738542556763, |
|
"eval_runtime": 28.5908, |
|
"eval_samples_per_second": 874.407, |
|
"eval_steps_per_second": 6.855, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.2798634812286689, |
|
"grad_norm": 7.232935905456543, |
|
"learning_rate": 1.7486569050879597e-05, |
|
"loss": 1.0235, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.3083048919226394, |
|
"grad_norm": 6.775105953216553, |
|
"learning_rate": 1.738122827346466e-05, |
|
"loss": 0.9949, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.3367463026166098, |
|
"grad_norm": 6.916153430938721, |
|
"learning_rate": 1.727588749604972e-05, |
|
"loss": 1.0058, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.36518771331058, |
|
"grad_norm": 6.561580181121826, |
|
"learning_rate": 1.7170546718634784e-05, |
|
"loss": 1.0039, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.36518771331058, |
|
"eval_loss": 1.1427565813064575, |
|
"eval_runtime": 28.6907, |
|
"eval_samples_per_second": 871.363, |
|
"eval_steps_per_second": 6.831, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.3936291240045506, |
|
"grad_norm": 6.508544921875, |
|
"learning_rate": 1.7065205941219847e-05, |
|
"loss": 1.0164, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.4220705346985212, |
|
"grad_norm": 7.889155387878418, |
|
"learning_rate": 1.695986516380491e-05, |
|
"loss": 0.9934, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.4505119453924915, |
|
"grad_norm": 7.1703782081604, |
|
"learning_rate": 1.685452438638997e-05, |
|
"loss": 0.9777, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.4789533560864618, |
|
"grad_norm": 7.198650360107422, |
|
"learning_rate": 1.6749183608975034e-05, |
|
"loss": 0.9753, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.4789533560864618, |
|
"eval_loss": 1.1101032495498657, |
|
"eval_runtime": 28.9361, |
|
"eval_samples_per_second": 863.971, |
|
"eval_steps_per_second": 6.774, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.5073947667804322, |
|
"grad_norm": 7.485228061676025, |
|
"learning_rate": 1.6643842831560098e-05, |
|
"loss": 0.9621, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 1.5358361774744027, |
|
"grad_norm": 6.426005840301514, |
|
"learning_rate": 1.653850205414516e-05, |
|
"loss": 0.9756, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.5642775881683733, |
|
"grad_norm": 6.803189277648926, |
|
"learning_rate": 1.643316127673022e-05, |
|
"loss": 0.9725, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 1.5927189988623436, |
|
"grad_norm": 7.307713508605957, |
|
"learning_rate": 1.6327820499315285e-05, |
|
"loss": 0.9649, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.5927189988623436, |
|
"eval_loss": 1.0812790393829346, |
|
"eval_runtime": 28.8811, |
|
"eval_samples_per_second": 865.619, |
|
"eval_steps_per_second": 6.786, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.621160409556314, |
|
"grad_norm": 6.56484317779541, |
|
"learning_rate": 1.6222479721900348e-05, |
|
"loss": 0.9652, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 1.6496018202502845, |
|
"grad_norm": 6.714264392852783, |
|
"learning_rate": 1.6117138944485412e-05, |
|
"loss": 0.9861, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.6780432309442548, |
|
"grad_norm": 6.9539642333984375, |
|
"learning_rate": 1.6011798167070475e-05, |
|
"loss": 0.916, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.7064846416382253, |
|
"grad_norm": 6.552751541137695, |
|
"learning_rate": 1.5906457389655535e-05, |
|
"loss": 0.9417, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.7064846416382253, |
|
"eval_loss": 1.0522855520248413, |
|
"eval_runtime": 28.864, |
|
"eval_samples_per_second": 866.132, |
|
"eval_steps_per_second": 6.79, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.7349260523321957, |
|
"grad_norm": 6.961670875549316, |
|
"learning_rate": 1.58011166122406e-05, |
|
"loss": 0.9599, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.763367463026166, |
|
"grad_norm": 7.874273300170898, |
|
"learning_rate": 1.5695775834825662e-05, |
|
"loss": 0.9275, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.7918088737201365, |
|
"grad_norm": 5.82428503036499, |
|
"learning_rate": 1.5590435057410726e-05, |
|
"loss": 0.9247, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.820250284414107, |
|
"grad_norm": 6.425380706787109, |
|
"learning_rate": 1.5485094279995786e-05, |
|
"loss": 0.9417, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.820250284414107, |
|
"eval_loss": 1.0305691957473755, |
|
"eval_runtime": 28.6406, |
|
"eval_samples_per_second": 872.888, |
|
"eval_steps_per_second": 6.843, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.8486916951080774, |
|
"grad_norm": 6.136819362640381, |
|
"learning_rate": 1.537975350258085e-05, |
|
"loss": 0.9275, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.8771331058020477, |
|
"grad_norm": 6.463824272155762, |
|
"learning_rate": 1.5274412725165913e-05, |
|
"loss": 0.9431, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.905574516496018, |
|
"grad_norm": 6.83174467086792, |
|
"learning_rate": 1.5169071947750974e-05, |
|
"loss": 0.9147, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.9340159271899886, |
|
"grad_norm": 7.504420280456543, |
|
"learning_rate": 1.5063731170336038e-05, |
|
"loss": 0.8957, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.9340159271899886, |
|
"eval_loss": 1.0050827264785767, |
|
"eval_runtime": 28.9461, |
|
"eval_samples_per_second": 863.675, |
|
"eval_steps_per_second": 6.771, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.9624573378839592, |
|
"grad_norm": 7.271299839019775, |
|
"learning_rate": 1.49583903929211e-05, |
|
"loss": 0.9169, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.9908987485779295, |
|
"grad_norm": 6.796669960021973, |
|
"learning_rate": 1.4853049615506163e-05, |
|
"loss": 0.9079, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.0193401592719, |
|
"grad_norm": 5.5628180503845215, |
|
"learning_rate": 1.4747708838091227e-05, |
|
"loss": 0.7057, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 2.04778156996587, |
|
"grad_norm": 5.777904987335205, |
|
"learning_rate": 1.4642368060676288e-05, |
|
"loss": 0.6037, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.04778156996587, |
|
"eval_loss": 0.9944195747375488, |
|
"eval_runtime": 28.8677, |
|
"eval_samples_per_second": 866.019, |
|
"eval_steps_per_second": 6.79, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.076222980659841, |
|
"grad_norm": 5.112311363220215, |
|
"learning_rate": 1.4537027283261352e-05, |
|
"loss": 0.5888, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 2.1046643913538112, |
|
"grad_norm": 6.392485618591309, |
|
"learning_rate": 1.4431686505846414e-05, |
|
"loss": 0.6134, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.1331058020477816, |
|
"grad_norm": 6.09423303604126, |
|
"learning_rate": 1.4326345728431477e-05, |
|
"loss": 0.6209, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 2.161547212741752, |
|
"grad_norm": 6.144412040710449, |
|
"learning_rate": 1.4221004951016539e-05, |
|
"loss": 0.6163, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.161547212741752, |
|
"eval_loss": 0.9836474061012268, |
|
"eval_runtime": 28.9354, |
|
"eval_samples_per_second": 863.993, |
|
"eval_steps_per_second": 6.774, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.189988623435722, |
|
"grad_norm": 5.410032272338867, |
|
"learning_rate": 1.4115664173601602e-05, |
|
"loss": 0.6271, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 2.218430034129693, |
|
"grad_norm": 5.688889980316162, |
|
"learning_rate": 1.4010323396186664e-05, |
|
"loss": 0.629, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.2468714448236633, |
|
"grad_norm": 5.400741100311279, |
|
"learning_rate": 1.3904982618771728e-05, |
|
"loss": 0.6041, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 2.2753128555176336, |
|
"grad_norm": 6.409387111663818, |
|
"learning_rate": 1.379964184135679e-05, |
|
"loss": 0.622, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.2753128555176336, |
|
"eval_loss": 0.9791940450668335, |
|
"eval_runtime": 29.3397, |
|
"eval_samples_per_second": 852.088, |
|
"eval_steps_per_second": 6.68, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.303754266211604, |
|
"grad_norm": 5.827444076538086, |
|
"learning_rate": 1.3694301063941853e-05, |
|
"loss": 0.6175, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 2.3321956769055747, |
|
"grad_norm": 6.436943054199219, |
|
"learning_rate": 1.3588960286526916e-05, |
|
"loss": 0.627, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 2.360637087599545, |
|
"grad_norm": 5.842226028442383, |
|
"learning_rate": 1.3483619509111978e-05, |
|
"loss": 0.6339, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 2.3890784982935154, |
|
"grad_norm": 6.457271575927734, |
|
"learning_rate": 1.3378278731697042e-05, |
|
"loss": 0.6325, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 2.3890784982935154, |
|
"eval_loss": 0.9643296003341675, |
|
"eval_runtime": 28.9755, |
|
"eval_samples_per_second": 862.799, |
|
"eval_steps_per_second": 6.764, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 2.4175199089874857, |
|
"grad_norm": 6.070743560791016, |
|
"learning_rate": 1.3272937954282103e-05, |
|
"loss": 0.6044, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 2.445961319681456, |
|
"grad_norm": 6.5427565574646, |
|
"learning_rate": 1.3167597176867167e-05, |
|
"loss": 0.6124, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 2.474402730375427, |
|
"grad_norm": 5.342416286468506, |
|
"learning_rate": 1.3062256399452229e-05, |
|
"loss": 0.6326, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 2.502844141069397, |
|
"grad_norm": 5.6298041343688965, |
|
"learning_rate": 1.2956915622037292e-05, |
|
"loss": 0.6349, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 2.502844141069397, |
|
"eval_loss": 0.9462358355522156, |
|
"eval_runtime": 29.0573, |
|
"eval_samples_per_second": 860.369, |
|
"eval_steps_per_second": 6.745, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 2.5312855517633674, |
|
"grad_norm": 5.618624210357666, |
|
"learning_rate": 1.2851574844622354e-05, |
|
"loss": 0.6286, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 2.5597269624573378, |
|
"grad_norm": 5.629756927490234, |
|
"learning_rate": 1.2746234067207417e-05, |
|
"loss": 0.6325, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 2.5881683731513085, |
|
"grad_norm": 5.6407318115234375, |
|
"learning_rate": 1.2640893289792479e-05, |
|
"loss": 0.6399, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 2.616609783845279, |
|
"grad_norm": 6.080498695373535, |
|
"learning_rate": 1.2535552512377542e-05, |
|
"loss": 0.6184, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 2.616609783845279, |
|
"eval_loss": 0.9317007064819336, |
|
"eval_runtime": 29.0538, |
|
"eval_samples_per_second": 860.472, |
|
"eval_steps_per_second": 6.746, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 2.645051194539249, |
|
"grad_norm": 6.4962239265441895, |
|
"learning_rate": 1.2430211734962604e-05, |
|
"loss": 0.6292, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 2.6734926052332195, |
|
"grad_norm": 6.621969223022461, |
|
"learning_rate": 1.2324870957547668e-05, |
|
"loss": 0.6017, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 2.70193401592719, |
|
"grad_norm": 5.2126054763793945, |
|
"learning_rate": 1.2219530180132731e-05, |
|
"loss": 0.6305, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 2.73037542662116, |
|
"grad_norm": 6.410334587097168, |
|
"learning_rate": 1.2114189402717793e-05, |
|
"loss": 0.6152, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 2.73037542662116, |
|
"eval_loss": 0.9212636947631836, |
|
"eval_runtime": 29.0224, |
|
"eval_samples_per_second": 861.404, |
|
"eval_steps_per_second": 6.753, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 2.758816837315131, |
|
"grad_norm": 6.005552291870117, |
|
"learning_rate": 1.2008848625302856e-05, |
|
"loss": 0.5972, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 2.7872582480091013, |
|
"grad_norm": 6.479732990264893, |
|
"learning_rate": 1.1903507847887918e-05, |
|
"loss": 0.6048, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 2.8156996587030716, |
|
"grad_norm": 6.2526397705078125, |
|
"learning_rate": 1.1798167070472982e-05, |
|
"loss": 0.6096, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 2.8441410693970424, |
|
"grad_norm": 6.823054313659668, |
|
"learning_rate": 1.1692826293058043e-05, |
|
"loss": 0.6156, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.8441410693970424, |
|
"eval_loss": 0.9072502851486206, |
|
"eval_runtime": 29.0918, |
|
"eval_samples_per_second": 859.348, |
|
"eval_steps_per_second": 6.737, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.8725824800910127, |
|
"grad_norm": 5.63970422744751, |
|
"learning_rate": 1.1587485515643107e-05, |
|
"loss": 0.5942, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 2.901023890784983, |
|
"grad_norm": 5.7269182205200195, |
|
"learning_rate": 1.1482144738228169e-05, |
|
"loss": 0.592, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 2.9294653014789533, |
|
"grad_norm": 6.235472202301025, |
|
"learning_rate": 1.1376803960813232e-05, |
|
"loss": 0.6088, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 2.9579067121729237, |
|
"grad_norm": 6.49041748046875, |
|
"learning_rate": 1.1271463183398294e-05, |
|
"loss": 0.5941, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 2.9579067121729237, |
|
"eval_loss": 0.8950417041778564, |
|
"eval_runtime": 29.0632, |
|
"eval_samples_per_second": 860.195, |
|
"eval_steps_per_second": 6.744, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 2.986348122866894, |
|
"grad_norm": 6.089723587036133, |
|
"learning_rate": 1.1166122405983357e-05, |
|
"loss": 0.6161, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 3.0147895335608648, |
|
"grad_norm": 4.977637767791748, |
|
"learning_rate": 1.1060781628568419e-05, |
|
"loss": 0.5021, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 3.043230944254835, |
|
"grad_norm": 5.729337215423584, |
|
"learning_rate": 1.0955440851153483e-05, |
|
"loss": 0.4116, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 3.0716723549488054, |
|
"grad_norm": 4.303124904632568, |
|
"learning_rate": 1.0850100073738546e-05, |
|
"loss": 0.3936, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 3.0716723549488054, |
|
"eval_loss": 0.9009103775024414, |
|
"eval_runtime": 28.839, |
|
"eval_samples_per_second": 866.881, |
|
"eval_steps_per_second": 6.796, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 3.1001137656427757, |
|
"grad_norm": 5.400048732757568, |
|
"learning_rate": 1.0744759296323608e-05, |
|
"loss": 0.4193, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 3.1285551763367465, |
|
"grad_norm": 6.018354415893555, |
|
"learning_rate": 1.0639418518908671e-05, |
|
"loss": 0.422, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 3.156996587030717, |
|
"grad_norm": 5.685466766357422, |
|
"learning_rate": 1.0534077741493733e-05, |
|
"loss": 0.432, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 3.185437997724687, |
|
"grad_norm": 5.172823905944824, |
|
"learning_rate": 1.0428736964078797e-05, |
|
"loss": 0.4281, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 3.185437997724687, |
|
"eval_loss": 0.8985010981559753, |
|
"eval_runtime": 28.8596, |
|
"eval_samples_per_second": 866.262, |
|
"eval_steps_per_second": 6.791, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 3.2138794084186575, |
|
"grad_norm": 4.836643218994141, |
|
"learning_rate": 1.0323396186663858e-05, |
|
"loss": 0.4091, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 3.242320819112628, |
|
"grad_norm": 5.528740406036377, |
|
"learning_rate": 1.0218055409248922e-05, |
|
"loss": 0.4305, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 3.2707622298065986, |
|
"grad_norm": 4.45158576965332, |
|
"learning_rate": 1.0112714631833984e-05, |
|
"loss": 0.4203, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 3.299203640500569, |
|
"grad_norm": 6.183067798614502, |
|
"learning_rate": 1.0007373854419047e-05, |
|
"loss": 0.4193, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 3.299203640500569, |
|
"eval_loss": 0.8869061470031738, |
|
"eval_runtime": 28.6962, |
|
"eval_samples_per_second": 871.197, |
|
"eval_steps_per_second": 6.83, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 3.3276450511945392, |
|
"grad_norm": 5.19403600692749, |
|
"learning_rate": 9.902033077004109e-06, |
|
"loss": 0.4238, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 3.3560864618885096, |
|
"grad_norm": 5.304056644439697, |
|
"learning_rate": 9.796692299589172e-06, |
|
"loss": 0.4274, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 3.3845278725824803, |
|
"grad_norm": 4.698873519897461, |
|
"learning_rate": 9.691351522174236e-06, |
|
"loss": 0.4124, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 3.4129692832764507, |
|
"grad_norm": 5.627292156219482, |
|
"learning_rate": 9.586010744759297e-06, |
|
"loss": 0.4241, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 3.4129692832764507, |
|
"eval_loss": 0.8842443823814392, |
|
"eval_runtime": 28.6817, |
|
"eval_samples_per_second": 871.636, |
|
"eval_steps_per_second": 6.834, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 3.441410693970421, |
|
"grad_norm": 6.473363876342773, |
|
"learning_rate": 9.480669967344361e-06, |
|
"loss": 0.427, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 3.4698521046643913, |
|
"grad_norm": 4.9653801918029785, |
|
"learning_rate": 9.375329189929423e-06, |
|
"loss": 0.4275, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 3.4982935153583616, |
|
"grad_norm": 4.9852294921875, |
|
"learning_rate": 9.269988412514486e-06, |
|
"loss": 0.4152, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 3.526734926052332, |
|
"grad_norm": 5.868428707122803, |
|
"learning_rate": 9.164647635099548e-06, |
|
"loss": 0.4247, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 3.526734926052332, |
|
"eval_loss": 0.8732792139053345, |
|
"eval_runtime": 28.8814, |
|
"eval_samples_per_second": 865.608, |
|
"eval_steps_per_second": 6.786, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 3.5551763367463027, |
|
"grad_norm": 5.333588600158691, |
|
"learning_rate": 9.05930685768461e-06, |
|
"loss": 0.4111, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 3.583617747440273, |
|
"grad_norm": 5.569532871246338, |
|
"learning_rate": 8.953966080269673e-06, |
|
"loss": 0.4396, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 3.6120591581342434, |
|
"grad_norm": 5.38419771194458, |
|
"learning_rate": 8.848625302854735e-06, |
|
"loss": 0.4122, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 3.640500568828214, |
|
"grad_norm": 5.328497409820557, |
|
"learning_rate": 8.743284525439798e-06, |
|
"loss": 0.4252, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 3.640500568828214, |
|
"eval_loss": 0.8656958937644958, |
|
"eval_runtime": 28.751, |
|
"eval_samples_per_second": 869.534, |
|
"eval_steps_per_second": 6.817, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 3.6689419795221845, |
|
"grad_norm": 5.675217151641846, |
|
"learning_rate": 8.63794374802486e-06, |
|
"loss": 0.4167, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 3.697383390216155, |
|
"grad_norm": 5.26973295211792, |
|
"learning_rate": 8.532602970609924e-06, |
|
"loss": 0.4282, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 3.725824800910125, |
|
"grad_norm": 5.991490840911865, |
|
"learning_rate": 8.427262193194985e-06, |
|
"loss": 0.411, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 3.7542662116040955, |
|
"grad_norm": 5.413957118988037, |
|
"learning_rate": 8.321921415780049e-06, |
|
"loss": 0.4273, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 3.7542662116040955, |
|
"eval_loss": 0.8539847135543823, |
|
"eval_runtime": 28.8669, |
|
"eval_samples_per_second": 866.045, |
|
"eval_steps_per_second": 6.79, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 3.782707622298066, |
|
"grad_norm": 5.672956466674805, |
|
"learning_rate": 8.21658063836511e-06, |
|
"loss": 0.4327, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 3.8111490329920366, |
|
"grad_norm": 6.0553059577941895, |
|
"learning_rate": 8.111239860950174e-06, |
|
"loss": 0.431, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 3.839590443686007, |
|
"grad_norm": 6.111351013183594, |
|
"learning_rate": 8.005899083535238e-06, |
|
"loss": 0.4347, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 3.868031854379977, |
|
"grad_norm": 6.185035705566406, |
|
"learning_rate": 7.9005583061203e-06, |
|
"loss": 0.4264, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 3.868031854379977, |
|
"eval_loss": 0.8523036241531372, |
|
"eval_runtime": 28.7415, |
|
"eval_samples_per_second": 869.823, |
|
"eval_steps_per_second": 6.819, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 3.8964732650739475, |
|
"grad_norm": 4.952618598937988, |
|
"learning_rate": 7.795217528705363e-06, |
|
"loss": 0.4213, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 3.9249146757679183, |
|
"grad_norm": 5.168086528778076, |
|
"learning_rate": 7.689876751290425e-06, |
|
"loss": 0.4285, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 3.9533560864618886, |
|
"grad_norm": 5.6217732429504395, |
|
"learning_rate": 7.584535973875487e-06, |
|
"loss": 0.4138, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 3.981797497155859, |
|
"grad_norm": 4.983550548553467, |
|
"learning_rate": 7.47919519646055e-06, |
|
"loss": 0.4051, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 3.981797497155859, |
|
"eval_loss": 0.8406953811645508, |
|
"eval_runtime": 28.8132, |
|
"eval_samples_per_second": 867.659, |
|
"eval_steps_per_second": 6.802, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 4.010238907849829, |
|
"grad_norm": 3.829274892807007, |
|
"learning_rate": 7.373854419045613e-06, |
|
"loss": 0.3779, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 4.0386803185438, |
|
"grad_norm": 4.154295921325684, |
|
"learning_rate": 7.268513641630676e-06, |
|
"loss": 0.2957, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 4.06712172923777, |
|
"grad_norm": 5.0097222328186035, |
|
"learning_rate": 7.1631728642157386e-06, |
|
"loss": 0.2939, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 4.09556313993174, |
|
"grad_norm": 5.015048027038574, |
|
"learning_rate": 7.057832086800801e-06, |
|
"loss": 0.3065, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 4.09556313993174, |
|
"eval_loss": 0.8590184450149536, |
|
"eval_runtime": 28.7607, |
|
"eval_samples_per_second": 869.241, |
|
"eval_steps_per_second": 6.815, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 4.1240045506257115, |
|
"grad_norm": 4.9901018142700195, |
|
"learning_rate": 6.952491309385864e-06, |
|
"loss": 0.3081, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 4.152445961319682, |
|
"grad_norm": 4.8424391746521, |
|
"learning_rate": 6.847150531970926e-06, |
|
"loss": 0.3043, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 4.180887372013652, |
|
"grad_norm": 5.147951602935791, |
|
"learning_rate": 6.741809754555989e-06, |
|
"loss": 0.3176, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 4.2093287827076225, |
|
"grad_norm": 4.292293548583984, |
|
"learning_rate": 6.636468977141052e-06, |
|
"loss": 0.3067, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 4.2093287827076225, |
|
"eval_loss": 0.848746657371521, |
|
"eval_runtime": 29.0524, |
|
"eval_samples_per_second": 860.514, |
|
"eval_steps_per_second": 6.746, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 4.237770193401593, |
|
"grad_norm": 4.796692848205566, |
|
"learning_rate": 6.531128199726114e-06, |
|
"loss": 0.299, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 4.266211604095563, |
|
"grad_norm": 5.196813583374023, |
|
"learning_rate": 6.425787422311177e-06, |
|
"loss": 0.3106, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 4.294653014789533, |
|
"grad_norm": 4.551479816436768, |
|
"learning_rate": 6.3204466448962395e-06, |
|
"loss": 0.3062, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 4.323094425483504, |
|
"grad_norm": 4.6921257972717285, |
|
"learning_rate": 6.215105867481302e-06, |
|
"loss": 0.3153, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 4.323094425483504, |
|
"eval_loss": 0.8497870564460754, |
|
"eval_runtime": 29.0027, |
|
"eval_samples_per_second": 861.988, |
|
"eval_steps_per_second": 6.758, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 4.351535836177474, |
|
"grad_norm": 4.535303592681885, |
|
"learning_rate": 6.109765090066366e-06, |
|
"loss": 0.3206, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 4.379977246871444, |
|
"grad_norm": 5.174567222595215, |
|
"learning_rate": 6.004424312651428e-06, |
|
"loss": 0.3202, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 4.408418657565416, |
|
"grad_norm": 4.402812480926514, |
|
"learning_rate": 5.899083535236491e-06, |
|
"loss": 0.3167, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 4.436860068259386, |
|
"grad_norm": 4.917297840118408, |
|
"learning_rate": 5.7937427578215534e-06, |
|
"loss": 0.3044, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 4.436860068259386, |
|
"eval_loss": 0.8426228165626526, |
|
"eval_runtime": 29.2233, |
|
"eval_samples_per_second": 855.482, |
|
"eval_steps_per_second": 6.707, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 4.465301478953356, |
|
"grad_norm": 5.476150989532471, |
|
"learning_rate": 5.688401980406616e-06, |
|
"loss": 0.3015, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 4.493742889647327, |
|
"grad_norm": 5.594091415405273, |
|
"learning_rate": 5.583061202991679e-06, |
|
"loss": 0.3157, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 4.522184300341297, |
|
"grad_norm": 4.798509120941162, |
|
"learning_rate": 5.477720425576741e-06, |
|
"loss": 0.3109, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 4.550625711035267, |
|
"grad_norm": 4.705766201019287, |
|
"learning_rate": 5.372379648161804e-06, |
|
"loss": 0.3164, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 4.550625711035267, |
|
"eval_loss": 0.8384647369384766, |
|
"eval_runtime": 29.0223, |
|
"eval_samples_per_second": 861.406, |
|
"eval_steps_per_second": 6.753, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 4.579067121729238, |
|
"grad_norm": 5.214234352111816, |
|
"learning_rate": 5.269145686295165e-06, |
|
"loss": 0.2996, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 4.607508532423208, |
|
"grad_norm": 3.9629294872283936, |
|
"learning_rate": 5.163804908880228e-06, |
|
"loss": 0.3247, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 4.635949943117178, |
|
"grad_norm": 5.35923957824707, |
|
"learning_rate": 5.058464131465291e-06, |
|
"loss": 0.3093, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 4.664391353811149, |
|
"grad_norm": 4.924727916717529, |
|
"learning_rate": 4.9531233540503534e-06, |
|
"loss": 0.3017, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 4.664391353811149, |
|
"eval_loss": 0.8293972611427307, |
|
"eval_runtime": 29.0332, |
|
"eval_samples_per_second": 861.084, |
|
"eval_steps_per_second": 6.751, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 4.69283276450512, |
|
"grad_norm": 4.929891586303711, |
|
"learning_rate": 4.847782576635416e-06, |
|
"loss": 0.3075, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 4.72127417519909, |
|
"grad_norm": 4.345849514007568, |
|
"learning_rate": 4.742441799220479e-06, |
|
"loss": 0.3006, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 4.74971558589306, |
|
"grad_norm": 4.58878231048584, |
|
"learning_rate": 4.637101021805541e-06, |
|
"loss": 0.3134, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 4.778156996587031, |
|
"grad_norm": 5.448882579803467, |
|
"learning_rate": 4.531760244390604e-06, |
|
"loss": 0.3111, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 4.778156996587031, |
|
"eval_loss": 0.8249350786209106, |
|
"eval_runtime": 29.1624, |
|
"eval_samples_per_second": 857.269, |
|
"eval_steps_per_second": 6.721, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 4.806598407281001, |
|
"grad_norm": 4.381404399871826, |
|
"learning_rate": 4.4264194669756665e-06, |
|
"loss": 0.3165, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 4.835039817974971, |
|
"grad_norm": 4.86619234085083, |
|
"learning_rate": 4.321078689560729e-06, |
|
"loss": 0.3071, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 4.863481228668942, |
|
"grad_norm": 5.313292503356934, |
|
"learning_rate": 4.215737912145792e-06, |
|
"loss": 0.3017, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 4.891922639362912, |
|
"grad_norm": 4.802574157714844, |
|
"learning_rate": 4.110397134730854e-06, |
|
"loss": 0.3092, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 4.891922639362912, |
|
"eval_loss": 0.8224520087242126, |
|
"eval_runtime": 29.0511, |
|
"eval_samples_per_second": 860.551, |
|
"eval_steps_per_second": 6.747, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 4.920364050056882, |
|
"grad_norm": 5.428598880767822, |
|
"learning_rate": 4.005056357315917e-06, |
|
"loss": 0.3, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 4.948805460750854, |
|
"grad_norm": 5.6783528327941895, |
|
"learning_rate": 3.8997155799009805e-06, |
|
"loss": 0.2999, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 4.977246871444824, |
|
"grad_norm": 5.2957940101623535, |
|
"learning_rate": 3.7943748024860427e-06, |
|
"loss": 0.3116, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 5.005688282138794, |
|
"grad_norm": 4.1276631355285645, |
|
"learning_rate": 3.6890340250711053e-06, |
|
"loss": 0.3046, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 5.005688282138794, |
|
"eval_loss": 0.8173409700393677, |
|
"eval_runtime": 28.9634, |
|
"eval_samples_per_second": 863.157, |
|
"eval_steps_per_second": 6.767, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 5.034129692832765, |
|
"grad_norm": 4.093660354614258, |
|
"learning_rate": 3.5836932476561683e-06, |
|
"loss": 0.2501, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 5.062571103526735, |
|
"grad_norm": 5.549435615539551, |
|
"learning_rate": 3.478352470241231e-06, |
|
"loss": 0.2443, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 5.091012514220705, |
|
"grad_norm": 4.558211803436279, |
|
"learning_rate": 3.3730116928262936e-06, |
|
"loss": 0.2338, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 5.1194539249146755, |
|
"grad_norm": 3.450760841369629, |
|
"learning_rate": 3.267670915411356e-06, |
|
"loss": 0.2382, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 5.1194539249146755, |
|
"eval_loss": 0.8248207569122314, |
|
"eval_runtime": 29.0514, |
|
"eval_samples_per_second": 860.545, |
|
"eval_steps_per_second": 6.747, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 5.147895335608646, |
|
"grad_norm": 4.0541205406188965, |
|
"learning_rate": 3.162330137996419e-06, |
|
"loss": 0.2524, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 5.176336746302616, |
|
"grad_norm": 4.376137733459473, |
|
"learning_rate": 3.0569893605814814e-06, |
|
"loss": 0.2427, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 5.204778156996587, |
|
"grad_norm": 4.169808864593506, |
|
"learning_rate": 2.951648583166544e-06, |
|
"loss": 0.2512, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 5.233219567690558, |
|
"grad_norm": 4.089740753173828, |
|
"learning_rate": 2.846307805751607e-06, |
|
"loss": 0.2377, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 5.233219567690558, |
|
"eval_loss": 0.8218184113502502, |
|
"eval_runtime": 28.9027, |
|
"eval_samples_per_second": 864.97, |
|
"eval_steps_per_second": 6.781, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 5.261660978384528, |
|
"grad_norm": 4.028066635131836, |
|
"learning_rate": 2.7409670283366697e-06, |
|
"loss": 0.2458, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 5.290102389078498, |
|
"grad_norm": 5.62259578704834, |
|
"learning_rate": 2.635626250921732e-06, |
|
"loss": 0.2515, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 5.318543799772469, |
|
"grad_norm": 4.931870937347412, |
|
"learning_rate": 2.5302854735067945e-06, |
|
"loss": 0.2453, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 5.346985210466439, |
|
"grad_norm": 4.307934284210205, |
|
"learning_rate": 2.4249446960918575e-06, |
|
"loss": 0.244, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 5.346985210466439, |
|
"eval_loss": 0.8225930333137512, |
|
"eval_runtime": 28.8011, |
|
"eval_samples_per_second": 868.022, |
|
"eval_steps_per_second": 6.805, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 5.375426621160409, |
|
"grad_norm": 3.650233030319214, |
|
"learning_rate": 2.31960391867692e-06, |
|
"loss": 0.2389, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 5.40386803185438, |
|
"grad_norm": 4.171177864074707, |
|
"learning_rate": 2.2142631412619828e-06, |
|
"loss": 0.253, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 5.43230944254835, |
|
"grad_norm": 5.055683135986328, |
|
"learning_rate": 2.1089223638470454e-06, |
|
"loss": 0.2509, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 5.460750853242321, |
|
"grad_norm": 4.621593952178955, |
|
"learning_rate": 2.003581586432108e-06, |
|
"loss": 0.2492, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 5.460750853242321, |
|
"eval_loss": 0.8198309540748596, |
|
"eval_runtime": 28.7042, |
|
"eval_samples_per_second": 870.954, |
|
"eval_steps_per_second": 6.828, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 5.489192263936292, |
|
"grad_norm": 5.461741924285889, |
|
"learning_rate": 1.8982408090171708e-06, |
|
"loss": 0.2379, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 5.517633674630262, |
|
"grad_norm": 4.083144664764404, |
|
"learning_rate": 1.7929000316022333e-06, |
|
"loss": 0.247, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 5.546075085324232, |
|
"grad_norm": 4.508319854736328, |
|
"learning_rate": 1.6875592541872959e-06, |
|
"loss": 0.2419, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 5.5745164960182025, |
|
"grad_norm": 4.420298099517822, |
|
"learning_rate": 1.5822184767723587e-06, |
|
"loss": 0.244, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 5.5745164960182025, |
|
"eval_loss": 0.8149560689926147, |
|
"eval_runtime": 28.7025, |
|
"eval_samples_per_second": 871.004, |
|
"eval_steps_per_second": 6.829, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 5.602957906712173, |
|
"grad_norm": 4.702558517456055, |
|
"learning_rate": 1.4768776993574213e-06, |
|
"loss": 0.2498, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 5.631399317406143, |
|
"grad_norm": 3.864471912384033, |
|
"learning_rate": 1.371536921942484e-06, |
|
"loss": 0.2381, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 5.6598407281001135, |
|
"grad_norm": 4.41420316696167, |
|
"learning_rate": 1.2661961445275468e-06, |
|
"loss": 0.2425, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 5.688282138794084, |
|
"grad_norm": 4.402945041656494, |
|
"learning_rate": 1.1608553671126094e-06, |
|
"loss": 0.2451, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 5.688282138794084, |
|
"eval_loss": 0.8147642016410828, |
|
"eval_runtime": 28.751, |
|
"eval_samples_per_second": 869.534, |
|
"eval_steps_per_second": 6.817, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 5.716723549488055, |
|
"grad_norm": 4.66687536239624, |
|
"learning_rate": 1.055514589697672e-06, |
|
"loss": 0.2468, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 5.745164960182025, |
|
"grad_norm": 4.6121649742126465, |
|
"learning_rate": 9.501738122827347e-07, |
|
"loss": 0.2404, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 5.773606370875996, |
|
"grad_norm": 4.210214614868164, |
|
"learning_rate": 8.469398504160961e-07, |
|
"loss": 0.2397, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 5.802047781569966, |
|
"grad_norm": 4.265695095062256, |
|
"learning_rate": 7.415990730011588e-07, |
|
"loss": 0.2417, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 5.802047781569966, |
|
"eval_loss": 0.8124446868896484, |
|
"eval_runtime": 28.7474, |
|
"eval_samples_per_second": 869.643, |
|
"eval_steps_per_second": 6.818, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 5.830489192263936, |
|
"grad_norm": 4.166738033294678, |
|
"learning_rate": 6.362582955862215e-07, |
|
"loss": 0.2446, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 5.858930602957907, |
|
"grad_norm": 4.40815544128418, |
|
"learning_rate": 5.309175181712841e-07, |
|
"loss": 0.2443, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 5.887372013651877, |
|
"grad_norm": 3.757612466812134, |
|
"learning_rate": 4.255767407563468e-07, |
|
"loss": 0.2465, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 5.915813424345847, |
|
"grad_norm": 5.059196472167969, |
|
"learning_rate": 3.202359633414095e-07, |
|
"loss": 0.2472, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 5.915813424345847, |
|
"eval_loss": 0.8121369481086731, |
|
"eval_runtime": 28.8178, |
|
"eval_samples_per_second": 867.521, |
|
"eval_steps_per_second": 6.801, |
|
"step": 10400 |
|
} |
|
], |
|
"logging_steps": 50, |
|
"max_steps": 10548, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 6, |
|
"save_steps": 200, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 128, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|