|
{ |
|
"best_metric": 3.619581912062131e-05, |
|
"best_model_checkpoint": "./bert_sensitive_columns/checkpoint-2200", |
|
"epoch": 10.0, |
|
"eval_steps": 500, |
|
"global_step": 2200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.045454545454545456, |
|
"grad_norm": 8.354640007019043, |
|
"learning_rate": 2.9863636363636365e-05, |
|
"loss": 0.6464, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.09090909090909091, |
|
"grad_norm": 4.68574857711792, |
|
"learning_rate": 2.972727272727273e-05, |
|
"loss": 0.6782, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.13636363636363635, |
|
"grad_norm": 18.99250602722168, |
|
"learning_rate": 2.959090909090909e-05, |
|
"loss": 0.5558, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.18181818181818182, |
|
"grad_norm": 6.562917709350586, |
|
"learning_rate": 2.9454545454545456e-05, |
|
"loss": 0.4202, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.22727272727272727, |
|
"grad_norm": 23.24289321899414, |
|
"learning_rate": 2.931818181818182e-05, |
|
"loss": 0.3803, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.2727272727272727, |
|
"grad_norm": 6.368680953979492, |
|
"learning_rate": 2.9181818181818185e-05, |
|
"loss": 0.436, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.3181818181818182, |
|
"grad_norm": 7.490790367126465, |
|
"learning_rate": 2.9045454545454546e-05, |
|
"loss": 0.359, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.36363636363636365, |
|
"grad_norm": 14.990336418151855, |
|
"learning_rate": 2.890909090909091e-05, |
|
"loss": 0.351, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.4090909090909091, |
|
"grad_norm": 16.408206939697266, |
|
"learning_rate": 2.8772727272727272e-05, |
|
"loss": 0.2892, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.45454545454545453, |
|
"grad_norm": 4.340272903442383, |
|
"learning_rate": 2.8636363636363637e-05, |
|
"loss": 0.3178, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 1.4607642889022827, |
|
"learning_rate": 2.8499999999999998e-05, |
|
"loss": 0.1817, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.5454545454545454, |
|
"grad_norm": 70.05758666992188, |
|
"learning_rate": 2.8363636363636363e-05, |
|
"loss": 0.1249, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5909090909090909, |
|
"grad_norm": 3.1135473251342773, |
|
"learning_rate": 2.8227272727272727e-05, |
|
"loss": 0.2626, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.6363636363636364, |
|
"grad_norm": 11.004677772521973, |
|
"learning_rate": 2.8090909090909092e-05, |
|
"loss": 0.2237, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6818181818181818, |
|
"grad_norm": 3.861924886703491, |
|
"learning_rate": 2.7954545454545453e-05, |
|
"loss": 0.1652, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.7272727272727273, |
|
"grad_norm": 0.520849883556366, |
|
"learning_rate": 2.7818181818181818e-05, |
|
"loss": 0.11, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.7727272727272727, |
|
"grad_norm": 0.3427947163581848, |
|
"learning_rate": 2.7681818181818183e-05, |
|
"loss": 0.0663, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.8181818181818182, |
|
"grad_norm": 2.3007407188415527, |
|
"learning_rate": 2.7545454545454547e-05, |
|
"loss": 0.302, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.8636363636363636, |
|
"grad_norm": 0.3552773594856262, |
|
"learning_rate": 2.7409090909090912e-05, |
|
"loss": 0.1245, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.9090909090909091, |
|
"grad_norm": 0.14678223431110382, |
|
"learning_rate": 2.7272727272727273e-05, |
|
"loss": 0.1086, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.9545454545454546, |
|
"grad_norm": 0.7700904607772827, |
|
"learning_rate": 2.7136363636363638e-05, |
|
"loss": 0.1189, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.1396052837371826, |
|
"learning_rate": 2.7000000000000002e-05, |
|
"loss": 0.1262, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.08161866664886475, |
|
"eval_runtime": 1.4104, |
|
"eval_samples_per_second": 622.523, |
|
"eval_steps_per_second": 38.996, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.0454545454545454, |
|
"grad_norm": 5.328164577484131, |
|
"learning_rate": 2.6863636363636367e-05, |
|
"loss": 0.0575, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.0909090909090908, |
|
"grad_norm": 0.3553941547870636, |
|
"learning_rate": 2.6727272727272728e-05, |
|
"loss": 0.1262, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.1363636363636362, |
|
"grad_norm": 0.1090177446603775, |
|
"learning_rate": 2.6590909090909093e-05, |
|
"loss": 0.0168, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.1818181818181819, |
|
"grad_norm": 3.0335617065429688, |
|
"learning_rate": 2.6454545454545454e-05, |
|
"loss": 0.0829, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.2272727272727273, |
|
"grad_norm": 40.902191162109375, |
|
"learning_rate": 2.631818181818182e-05, |
|
"loss": 0.0295, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.2727272727272727, |
|
"grad_norm": 0.07629093527793884, |
|
"learning_rate": 2.618181818181818e-05, |
|
"loss": 0.0022, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.3181818181818181, |
|
"grad_norm": 9.37717056274414, |
|
"learning_rate": 2.6045454545454545e-05, |
|
"loss": 0.1626, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.3636363636363638, |
|
"grad_norm": 0.09757604449987411, |
|
"learning_rate": 2.590909090909091e-05, |
|
"loss": 0.0302, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.4090909090909092, |
|
"grad_norm": 2.172060966491699, |
|
"learning_rate": 2.5772727272727274e-05, |
|
"loss": 0.0301, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.4545454545454546, |
|
"grad_norm": 0.06304845958948135, |
|
"learning_rate": 2.5636363636363635e-05, |
|
"loss": 0.0411, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.07031694054603577, |
|
"learning_rate": 2.55e-05, |
|
"loss": 0.0139, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.5454545454545454, |
|
"grad_norm": 3.3001744747161865, |
|
"learning_rate": 2.5363636363636364e-05, |
|
"loss": 0.1424, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.5909090909090908, |
|
"grad_norm": 0.08083586394786835, |
|
"learning_rate": 2.522727272727273e-05, |
|
"loss": 0.0291, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.6363636363636362, |
|
"grad_norm": 34.89284133911133, |
|
"learning_rate": 2.509090909090909e-05, |
|
"loss": 0.0757, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.6818181818181817, |
|
"grad_norm": 0.082089863717556, |
|
"learning_rate": 2.4954545454545455e-05, |
|
"loss": 0.0143, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.7272727272727273, |
|
"grad_norm": 0.2883528172969818, |
|
"learning_rate": 2.481818181818182e-05, |
|
"loss": 0.0769, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.7727272727272727, |
|
"grad_norm": 11.552708625793457, |
|
"learning_rate": 2.4681818181818184e-05, |
|
"loss": 0.0487, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.8181818181818183, |
|
"grad_norm": 0.09678918868303299, |
|
"learning_rate": 2.454545454545455e-05, |
|
"loss": 0.0612, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.8636363636363638, |
|
"grad_norm": 0.08423396944999695, |
|
"learning_rate": 2.440909090909091e-05, |
|
"loss": 0.0214, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.9090909090909092, |
|
"grad_norm": 97.11893463134766, |
|
"learning_rate": 2.4272727272727275e-05, |
|
"loss": 0.1287, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.9545454545454546, |
|
"grad_norm": 0.04146511107683182, |
|
"learning_rate": 2.4136363636363636e-05, |
|
"loss": 0.0202, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.03929423540830612, |
|
"learning_rate": 2.4e-05, |
|
"loss": 0.0047, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.010260566137731075, |
|
"eval_runtime": 0.9845, |
|
"eval_samples_per_second": 891.815, |
|
"eval_steps_per_second": 55.865, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.0454545454545454, |
|
"grad_norm": 0.10950633883476257, |
|
"learning_rate": 2.3863636363636362e-05, |
|
"loss": 0.0509, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.090909090909091, |
|
"grad_norm": 0.02540852129459381, |
|
"learning_rate": 2.3727272727272726e-05, |
|
"loss": 0.0476, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.1363636363636362, |
|
"grad_norm": 0.024377569556236267, |
|
"learning_rate": 2.359090909090909e-05, |
|
"loss": 0.0356, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.1818181818181817, |
|
"grad_norm": 0.025098495185375214, |
|
"learning_rate": 2.3454545454545456e-05, |
|
"loss": 0.0009, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.227272727272727, |
|
"grad_norm": 0.06944375485181808, |
|
"learning_rate": 2.3318181818181817e-05, |
|
"loss": 0.0193, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.2727272727272725, |
|
"grad_norm": 28.133596420288086, |
|
"learning_rate": 2.318181818181818e-05, |
|
"loss": 0.0339, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.3181818181818183, |
|
"grad_norm": 0.025236543267965317, |
|
"learning_rate": 2.3045454545454546e-05, |
|
"loss": 0.0462, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.3636363636363638, |
|
"grad_norm": 0.022095683962106705, |
|
"learning_rate": 2.290909090909091e-05, |
|
"loss": 0.0248, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.409090909090909, |
|
"grad_norm": 0.02965674363076687, |
|
"learning_rate": 2.2772727272727272e-05, |
|
"loss": 0.0023, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.4545454545454546, |
|
"grad_norm": 0.02505609020590782, |
|
"learning_rate": 2.2636363636363637e-05, |
|
"loss": 0.031, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.0254733357578516, |
|
"learning_rate": 2.25e-05, |
|
"loss": 0.0019, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.5454545454545454, |
|
"grad_norm": 0.018397022038698196, |
|
"learning_rate": 2.2363636363636366e-05, |
|
"loss": 0.0006, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.590909090909091, |
|
"grad_norm": 0.01716865971684456, |
|
"learning_rate": 2.222727272727273e-05, |
|
"loss": 0.0517, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.6363636363636362, |
|
"grad_norm": 0.016052110120654106, |
|
"learning_rate": 2.2090909090909092e-05, |
|
"loss": 0.0005, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.6818181818181817, |
|
"grad_norm": 0.025269588455557823, |
|
"learning_rate": 2.1954545454545457e-05, |
|
"loss": 0.002, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.7272727272727275, |
|
"grad_norm": 0.04568961635231972, |
|
"learning_rate": 2.1818181818181818e-05, |
|
"loss": 0.036, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.7727272727272725, |
|
"grad_norm": 0.01690821908414364, |
|
"learning_rate": 2.1681818181818182e-05, |
|
"loss": 0.0239, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.8181818181818183, |
|
"grad_norm": 0.02976076677441597, |
|
"learning_rate": 2.1545454545454544e-05, |
|
"loss": 0.0005, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.8636363636363638, |
|
"grad_norm": 0.15770655870437622, |
|
"learning_rate": 2.140909090909091e-05, |
|
"loss": 0.0006, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.909090909090909, |
|
"grad_norm": 0.011741632595658302, |
|
"learning_rate": 2.1272727272727273e-05, |
|
"loss": 0.0004, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.9545454545454546, |
|
"grad_norm": 0.012714399956166744, |
|
"learning_rate": 2.1136363636363638e-05, |
|
"loss": 0.0004, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.016626961529254913, |
|
"learning_rate": 2.1e-05, |
|
"loss": 0.0383, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.0007602364639751613, |
|
"eval_runtime": 1.0081, |
|
"eval_samples_per_second": 870.902, |
|
"eval_steps_per_second": 54.555, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 3.0454545454545454, |
|
"grad_norm": 0.010502061806619167, |
|
"learning_rate": 2.0863636363636363e-05, |
|
"loss": 0.0004, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 3.090909090909091, |
|
"grad_norm": 0.011059875600039959, |
|
"learning_rate": 2.0727272727272728e-05, |
|
"loss": 0.0003, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 3.1363636363636362, |
|
"grad_norm": 0.013830793090164661, |
|
"learning_rate": 2.0590909090909093e-05, |
|
"loss": 0.0005, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 3.1818181818181817, |
|
"grad_norm": 0.016489654779434204, |
|
"learning_rate": 2.0454545454545454e-05, |
|
"loss": 0.0014, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 3.227272727272727, |
|
"grad_norm": 0.011767825111746788, |
|
"learning_rate": 2.031818181818182e-05, |
|
"loss": 0.0004, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 3.2727272727272725, |
|
"grad_norm": 0.01013511698693037, |
|
"learning_rate": 2.0181818181818183e-05, |
|
"loss": 0.0003, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 3.3181818181818183, |
|
"grad_norm": 0.016494890674948692, |
|
"learning_rate": 2.0045454545454548e-05, |
|
"loss": 0.0941, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 3.3636363636363638, |
|
"grad_norm": 0.028399920091032982, |
|
"learning_rate": 1.9909090909090913e-05, |
|
"loss": 0.0003, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 3.409090909090909, |
|
"grad_norm": 0.006911724805831909, |
|
"learning_rate": 1.9772727272727274e-05, |
|
"loss": 0.0003, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 3.4545454545454546, |
|
"grad_norm": 0.009757892228662968, |
|
"learning_rate": 1.963636363636364e-05, |
|
"loss": 0.0003, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 0.014759697020053864, |
|
"learning_rate": 1.95e-05, |
|
"loss": 0.0702, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 3.5454545454545454, |
|
"grad_norm": 0.018820617347955704, |
|
"learning_rate": 1.9363636363636364e-05, |
|
"loss": 0.0023, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 3.590909090909091, |
|
"grad_norm": 0.008335668593645096, |
|
"learning_rate": 1.9227272727272726e-05, |
|
"loss": 0.099, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 3.6363636363636362, |
|
"grad_norm": 0.007782892789691687, |
|
"learning_rate": 1.909090909090909e-05, |
|
"loss": 0.0005, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 3.6818181818181817, |
|
"grad_norm": 0.007866962812840939, |
|
"learning_rate": 1.8954545454545455e-05, |
|
"loss": 0.0007, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 3.7272727272727275, |
|
"grad_norm": 0.00743032805621624, |
|
"learning_rate": 1.881818181818182e-05, |
|
"loss": 0.0003, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 3.7727272727272725, |
|
"grad_norm": 0.007409967016428709, |
|
"learning_rate": 1.868181818181818e-05, |
|
"loss": 0.0128, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 3.8181818181818183, |
|
"grad_norm": 0.008645043708384037, |
|
"learning_rate": 1.8545454545454545e-05, |
|
"loss": 0.0002, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 3.8636363636363638, |
|
"grad_norm": 0.006393834948539734, |
|
"learning_rate": 1.840909090909091e-05, |
|
"loss": 0.0002, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 3.909090909090909, |
|
"grad_norm": 0.008199839852750301, |
|
"learning_rate": 1.8272727272727275e-05, |
|
"loss": 0.0003, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 3.9545454545454546, |
|
"grad_norm": 0.007581517565995455, |
|
"learning_rate": 1.8136363636363636e-05, |
|
"loss": 0.0002, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 0.007007090840488672, |
|
"learning_rate": 1.8e-05, |
|
"loss": 0.0002, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.00015143574273679405, |
|
"eval_runtime": 0.9926, |
|
"eval_samples_per_second": 884.547, |
|
"eval_steps_per_second": 55.41, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 4.045454545454546, |
|
"grad_norm": 81.15371704101562, |
|
"learning_rate": 1.7863636363636365e-05, |
|
"loss": 0.006, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 4.090909090909091, |
|
"grad_norm": 0.01149928942322731, |
|
"learning_rate": 1.772727272727273e-05, |
|
"loss": 0.0002, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 4.136363636363637, |
|
"grad_norm": 0.0050843264907598495, |
|
"learning_rate": 1.759090909090909e-05, |
|
"loss": 0.0002, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 4.181818181818182, |
|
"grad_norm": 0.005241791717708111, |
|
"learning_rate": 1.7454545454545456e-05, |
|
"loss": 0.0002, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 4.2272727272727275, |
|
"grad_norm": 0.004743785131722689, |
|
"learning_rate": 1.731818181818182e-05, |
|
"loss": 0.0002, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 4.2727272727272725, |
|
"grad_norm": 0.004982436075806618, |
|
"learning_rate": 1.718181818181818e-05, |
|
"loss": 0.0002, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 4.318181818181818, |
|
"grad_norm": 0.005252942908555269, |
|
"learning_rate": 1.7045454545454546e-05, |
|
"loss": 0.0002, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 4.363636363636363, |
|
"grad_norm": 0.00490264967083931, |
|
"learning_rate": 1.6909090909090907e-05, |
|
"loss": 0.0002, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 4.409090909090909, |
|
"grad_norm": 0.003635741537436843, |
|
"learning_rate": 1.6772727272727272e-05, |
|
"loss": 0.0001, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 4.454545454545454, |
|
"grad_norm": 0.005204927641898394, |
|
"learning_rate": 1.6636363636363637e-05, |
|
"loss": 0.0001, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"grad_norm": 0.009996837005019188, |
|
"learning_rate": 1.65e-05, |
|
"loss": 0.0002, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 4.545454545454545, |
|
"grad_norm": 0.017751624807715416, |
|
"learning_rate": 1.6363636363636363e-05, |
|
"loss": 0.0315, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 4.590909090909091, |
|
"grad_norm": 0.00822280440479517, |
|
"learning_rate": 1.6227272727272727e-05, |
|
"loss": 0.0001, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 4.636363636363637, |
|
"grad_norm": 0.005263584200292826, |
|
"learning_rate": 1.6090909090909092e-05, |
|
"loss": 0.0001, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 4.681818181818182, |
|
"grad_norm": 0.004647717345505953, |
|
"learning_rate": 1.5954545454545456e-05, |
|
"loss": 0.0002, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 4.7272727272727275, |
|
"grad_norm": 0.0030076594557613134, |
|
"learning_rate": 1.5818181818181818e-05, |
|
"loss": 0.0001, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 4.7727272727272725, |
|
"grad_norm": 0.0052589308470487595, |
|
"learning_rate": 1.5681818181818182e-05, |
|
"loss": 0.0001, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 4.818181818181818, |
|
"grad_norm": 0.003165638307109475, |
|
"learning_rate": 1.5545454545454547e-05, |
|
"loss": 0.0001, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 4.863636363636363, |
|
"grad_norm": 0.0037285718135535717, |
|
"learning_rate": 1.540909090909091e-05, |
|
"loss": 0.0001, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 4.909090909090909, |
|
"grad_norm": 0.004402661230415106, |
|
"learning_rate": 1.5272727272727273e-05, |
|
"loss": 0.0001, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 4.954545454545455, |
|
"grad_norm": 0.003948619589209557, |
|
"learning_rate": 1.5136363636363636e-05, |
|
"loss": 0.0001, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 0.0029597911052405834, |
|
"learning_rate": 1.5e-05, |
|
"loss": 0.0001, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 8.946753951022401e-05, |
|
"eval_runtime": 0.9926, |
|
"eval_samples_per_second": 884.51, |
|
"eval_steps_per_second": 55.408, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 5.045454545454546, |
|
"grad_norm": 0.00478377053514123, |
|
"learning_rate": 1.4863636363636365e-05, |
|
"loss": 0.0001, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 5.090909090909091, |
|
"grad_norm": 0.0036518580745905638, |
|
"learning_rate": 1.4727272727272728e-05, |
|
"loss": 0.0001, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 5.136363636363637, |
|
"grad_norm": 0.004087444860488176, |
|
"learning_rate": 1.4590909090909093e-05, |
|
"loss": 0.0001, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 5.181818181818182, |
|
"grad_norm": 0.004169765394181013, |
|
"learning_rate": 1.4454545454545456e-05, |
|
"loss": 0.0001, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 5.2272727272727275, |
|
"grad_norm": 0.00350973685272038, |
|
"learning_rate": 1.4318181818181818e-05, |
|
"loss": 0.0001, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 5.2727272727272725, |
|
"grad_norm": 0.0037286856677383184, |
|
"learning_rate": 1.4181818181818181e-05, |
|
"loss": 0.0001, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 5.318181818181818, |
|
"grad_norm": 0.003630703780800104, |
|
"learning_rate": 1.4045454545454546e-05, |
|
"loss": 0.0001, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 5.363636363636363, |
|
"grad_norm": 0.0036072884686291218, |
|
"learning_rate": 1.3909090909090909e-05, |
|
"loss": 0.0001, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 5.409090909090909, |
|
"grad_norm": 0.004187653306871653, |
|
"learning_rate": 1.3772727272727274e-05, |
|
"loss": 0.0001, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 5.454545454545454, |
|
"grad_norm": 0.00530035886913538, |
|
"learning_rate": 1.3636363636363637e-05, |
|
"loss": 0.0001, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"grad_norm": 0.0027139252051711082, |
|
"learning_rate": 1.3500000000000001e-05, |
|
"loss": 0.0001, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 5.545454545454545, |
|
"grad_norm": 0.0034846195485442877, |
|
"learning_rate": 1.3363636363636364e-05, |
|
"loss": 0.0001, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 5.590909090909091, |
|
"grad_norm": 0.003408952383324504, |
|
"learning_rate": 1.3227272727272727e-05, |
|
"loss": 0.0001, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 5.636363636363637, |
|
"grad_norm": 0.0027936245314776897, |
|
"learning_rate": 1.309090909090909e-05, |
|
"loss": 0.0001, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 5.681818181818182, |
|
"grad_norm": 0.002841574139893055, |
|
"learning_rate": 1.2954545454545455e-05, |
|
"loss": 0.0001, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 5.7272727272727275, |
|
"grad_norm": 0.26475799083709717, |
|
"learning_rate": 1.2818181818181818e-05, |
|
"loss": 0.0001, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 5.7727272727272725, |
|
"grad_norm": 0.0034859974402934313, |
|
"learning_rate": 1.2681818181818182e-05, |
|
"loss": 0.0001, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 5.818181818181818, |
|
"grad_norm": 0.0028127585537731647, |
|
"learning_rate": 1.2545454545454545e-05, |
|
"loss": 0.0001, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 5.863636363636363, |
|
"grad_norm": 0.002384282648563385, |
|
"learning_rate": 1.240909090909091e-05, |
|
"loss": 0.0001, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 5.909090909090909, |
|
"grad_norm": 0.0030948910862207413, |
|
"learning_rate": 1.2272727272727274e-05, |
|
"loss": 0.0001, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 5.954545454545455, |
|
"grad_norm": 0.002442040015012026, |
|
"learning_rate": 1.2136363636363637e-05, |
|
"loss": 0.0001, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 0.0018502968596294522, |
|
"learning_rate": 1.2e-05, |
|
"loss": 0.0001, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 6.235863111214712e-05, |
|
"eval_runtime": 0.9927, |
|
"eval_samples_per_second": 884.48, |
|
"eval_steps_per_second": 55.406, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 6.045454545454546, |
|
"grad_norm": 0.002548688091337681, |
|
"learning_rate": 1.1863636363636363e-05, |
|
"loss": 0.0001, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 6.090909090909091, |
|
"grad_norm": 0.002689023967832327, |
|
"learning_rate": 1.1727272727272728e-05, |
|
"loss": 0.0001, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 6.136363636363637, |
|
"grad_norm": 0.002400546334683895, |
|
"learning_rate": 1.159090909090909e-05, |
|
"loss": 0.0001, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 6.181818181818182, |
|
"grad_norm": 0.0029753774870187044, |
|
"learning_rate": 1.1454545454545455e-05, |
|
"loss": 0.0001, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 6.2272727272727275, |
|
"grad_norm": 0.0020764051005244255, |
|
"learning_rate": 1.1318181818181818e-05, |
|
"loss": 0.0001, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 6.2727272727272725, |
|
"grad_norm": 0.00242880592122674, |
|
"learning_rate": 1.1181818181818183e-05, |
|
"loss": 0.0004, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 6.318181818181818, |
|
"grad_norm": 0.013365722261369228, |
|
"learning_rate": 1.1045454545454546e-05, |
|
"loss": 0.0001, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 6.363636363636363, |
|
"grad_norm": 0.0019247201271355152, |
|
"learning_rate": 1.0909090909090909e-05, |
|
"loss": 0.0001, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 6.409090909090909, |
|
"grad_norm": 0.0019971681758761406, |
|
"learning_rate": 1.0772727272727272e-05, |
|
"loss": 0.0001, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 6.454545454545454, |
|
"grad_norm": 0.002282345201820135, |
|
"learning_rate": 1.0636363636363636e-05, |
|
"loss": 0.0001, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"grad_norm": 0.0025554117746651173, |
|
"learning_rate": 1.05e-05, |
|
"loss": 0.0001, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 6.545454545454545, |
|
"grad_norm": 0.003260772442445159, |
|
"learning_rate": 1.0363636363636364e-05, |
|
"loss": 0.0001, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 6.590909090909091, |
|
"grad_norm": 0.0019251167541369796, |
|
"learning_rate": 1.0227272727272727e-05, |
|
"loss": 0.0001, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 6.636363636363637, |
|
"grad_norm": 0.002389734610915184, |
|
"learning_rate": 1.0090909090909092e-05, |
|
"loss": 0.0001, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 6.681818181818182, |
|
"grad_norm": 0.00212781666778028, |
|
"learning_rate": 9.954545454545456e-06, |
|
"loss": 0.0001, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 6.7272727272727275, |
|
"grad_norm": 0.003619167488068342, |
|
"learning_rate": 9.81818181818182e-06, |
|
"loss": 0.0001, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 6.7727272727272725, |
|
"grad_norm": 0.002498344052582979, |
|
"learning_rate": 9.681818181818182e-06, |
|
"loss": 0.0001, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 6.818181818181818, |
|
"grad_norm": 0.0023231736849993467, |
|
"learning_rate": 9.545454545454545e-06, |
|
"loss": 0.0001, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 6.863636363636363, |
|
"grad_norm": 0.0021667128894478083, |
|
"learning_rate": 9.40909090909091e-06, |
|
"loss": 0.003, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 6.909090909090909, |
|
"grad_norm": 0.0022279066033661366, |
|
"learning_rate": 9.272727272727273e-06, |
|
"loss": 0.0001, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 6.954545454545455, |
|
"grad_norm": 0.00177993334364146, |
|
"learning_rate": 9.136363636363637e-06, |
|
"loss": 0.0001, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"grad_norm": 0.0025669343303889036, |
|
"learning_rate": 9e-06, |
|
"loss": 0.0001, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 4.8654597776476294e-05, |
|
"eval_runtime": 0.981, |
|
"eval_samples_per_second": 895.028, |
|
"eval_steps_per_second": 56.067, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 7.045454545454546, |
|
"grad_norm": 0.0023670855443924665, |
|
"learning_rate": 8.863636363636365e-06, |
|
"loss": 0.0001, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 7.090909090909091, |
|
"grad_norm": 0.0023604007437825203, |
|
"learning_rate": 8.727272727272728e-06, |
|
"loss": 0.0001, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 7.136363636363637, |
|
"grad_norm": 0.0018116602441295981, |
|
"learning_rate": 8.59090909090909e-06, |
|
"loss": 0.0001, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 7.181818181818182, |
|
"grad_norm": 0.00249605649150908, |
|
"learning_rate": 8.454545454545454e-06, |
|
"loss": 0.0001, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 7.2272727272727275, |
|
"grad_norm": 0.001895196153782308, |
|
"learning_rate": 8.318181818181818e-06, |
|
"loss": 0.0001, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 7.2727272727272725, |
|
"grad_norm": 0.0017933849012479186, |
|
"learning_rate": 8.181818181818181e-06, |
|
"loss": 0.0429, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 7.318181818181818, |
|
"grad_norm": 0.0016959038330242038, |
|
"learning_rate": 8.045454545454546e-06, |
|
"loss": 0.0001, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 7.363636363636363, |
|
"grad_norm": 0.0016535187605768442, |
|
"learning_rate": 7.909090909090909e-06, |
|
"loss": 0.0001, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 7.409090909090909, |
|
"grad_norm": 0.0020366287790238857, |
|
"learning_rate": 7.772727272727273e-06, |
|
"loss": 0.0001, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 7.454545454545454, |
|
"grad_norm": 0.0017039361409842968, |
|
"learning_rate": 7.636363636363636e-06, |
|
"loss": 0.0001, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"grad_norm": 0.0018439262639731169, |
|
"learning_rate": 7.5e-06, |
|
"loss": 0.0001, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 7.545454545454545, |
|
"grad_norm": 0.0017576288664713502, |
|
"learning_rate": 7.363636363636364e-06, |
|
"loss": 0.0001, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 7.590909090909091, |
|
"grad_norm": 0.0015827094903215766, |
|
"learning_rate": 7.227272727272728e-06, |
|
"loss": 0.0001, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 7.636363636363637, |
|
"grad_norm": 0.0021957652643322945, |
|
"learning_rate": 7.090909090909091e-06, |
|
"loss": 0.0001, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 7.681818181818182, |
|
"grad_norm": 0.0018625753000378609, |
|
"learning_rate": 6.9545454545454545e-06, |
|
"loss": 0.0001, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 7.7272727272727275, |
|
"grad_norm": 0.003254745388403535, |
|
"learning_rate": 6.818181818181818e-06, |
|
"loss": 0.0001, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 7.7727272727272725, |
|
"grad_norm": 0.0016077288892120123, |
|
"learning_rate": 6.681818181818182e-06, |
|
"loss": 0.0001, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 7.818181818181818, |
|
"grad_norm": 0.0023993789218366146, |
|
"learning_rate": 6.545454545454545e-06, |
|
"loss": 0.0001, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 7.863636363636363, |
|
"grad_norm": 0.001893221982754767, |
|
"learning_rate": 6.409090909090909e-06, |
|
"loss": 0.0001, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 7.909090909090909, |
|
"grad_norm": 0.0018822376150637865, |
|
"learning_rate": 6.272727272727273e-06, |
|
"loss": 0.0001, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 7.954545454545455, |
|
"grad_norm": 0.009597906842827797, |
|
"learning_rate": 6.136363636363637e-06, |
|
"loss": 0.0001, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 0.0014578086556866765, |
|
"learning_rate": 6e-06, |
|
"loss": 0.0001, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 4.155210262979381e-05, |
|
"eval_runtime": 1.017, |
|
"eval_samples_per_second": 863.319, |
|
"eval_steps_per_second": 54.08, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 8.045454545454545, |
|
"grad_norm": 0.001562977209687233, |
|
"learning_rate": 5.863636363636364e-06, |
|
"loss": 0.0001, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 8.090909090909092, |
|
"grad_norm": 0.0019722983706742525, |
|
"learning_rate": 5.727272727272728e-06, |
|
"loss": 0.0001, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 8.136363636363637, |
|
"grad_norm": 0.0016468315152451396, |
|
"learning_rate": 5.5909090909090915e-06, |
|
"loss": 0.0001, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 8.181818181818182, |
|
"grad_norm": 0.0018590294057503343, |
|
"learning_rate": 5.4545454545454545e-06, |
|
"loss": 0.0001, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 8.227272727272727, |
|
"grad_norm": 0.002041436033323407, |
|
"learning_rate": 5.318181818181818e-06, |
|
"loss": 0.0001, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 8.272727272727273, |
|
"grad_norm": 0.0021510140504688025, |
|
"learning_rate": 5.181818181818182e-06, |
|
"loss": 0.0001, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 8.318181818181818, |
|
"grad_norm": 0.0015175098087638617, |
|
"learning_rate": 5.045454545454546e-06, |
|
"loss": 0.0001, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 8.363636363636363, |
|
"grad_norm": 0.001754813943989575, |
|
"learning_rate": 4.90909090909091e-06, |
|
"loss": 0.0001, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 8.409090909090908, |
|
"grad_norm": 0.001608902239240706, |
|
"learning_rate": 4.7727272727272725e-06, |
|
"loss": 0.0001, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 8.454545454545455, |
|
"grad_norm": 0.002168968552723527, |
|
"learning_rate": 4.636363636363636e-06, |
|
"loss": 0.0001, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"grad_norm": 0.001456632511690259, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.0, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 8.545454545454545, |
|
"grad_norm": 0.0017024242551997304, |
|
"learning_rate": 4.363636363636364e-06, |
|
"loss": 0.0001, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 8.590909090909092, |
|
"grad_norm": 0.00176154519431293, |
|
"learning_rate": 4.227272727272727e-06, |
|
"loss": 0.0001, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 8.636363636363637, |
|
"grad_norm": 0.0019339303253218532, |
|
"learning_rate": 4.090909090909091e-06, |
|
"loss": 0.0001, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 8.681818181818182, |
|
"grad_norm": 0.0019142951350659132, |
|
"learning_rate": 3.954545454545454e-06, |
|
"loss": 0.0001, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 8.727272727272727, |
|
"grad_norm": 0.0015304730040952563, |
|
"learning_rate": 3.818181818181818e-06, |
|
"loss": 0.0001, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 8.772727272727273, |
|
"grad_norm": 0.0021663594525307417, |
|
"learning_rate": 3.681818181818182e-06, |
|
"loss": 0.0001, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 8.818181818181818, |
|
"grad_norm": 0.001586704864166677, |
|
"learning_rate": 3.5454545454545454e-06, |
|
"loss": 0.0001, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 8.863636363636363, |
|
"grad_norm": 0.001986338524147868, |
|
"learning_rate": 3.409090909090909e-06, |
|
"loss": 0.0, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 8.909090909090908, |
|
"grad_norm": 0.001552366535179317, |
|
"learning_rate": 3.2727272727272725e-06, |
|
"loss": 0.0033, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 8.954545454545455, |
|
"grad_norm": 0.0018597301095724106, |
|
"learning_rate": 3.1363636363636363e-06, |
|
"loss": 0.0001, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"grad_norm": 0.002290609758347273, |
|
"learning_rate": 3e-06, |
|
"loss": 0.0, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 3.7486017390619963e-05, |
|
"eval_runtime": 1.0378, |
|
"eval_samples_per_second": 846.047, |
|
"eval_steps_per_second": 52.998, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 9.045454545454545, |
|
"grad_norm": 0.0014905119314789772, |
|
"learning_rate": 2.863636363636364e-06, |
|
"loss": 0.0, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 9.090909090909092, |
|
"grad_norm": 0.0013228630414232612, |
|
"learning_rate": 2.7272727272727272e-06, |
|
"loss": 0.0001, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 9.136363636363637, |
|
"grad_norm": 0.00189464061986655, |
|
"learning_rate": 2.590909090909091e-06, |
|
"loss": 0.0, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 9.181818181818182, |
|
"grad_norm": 0.0020432292949408293, |
|
"learning_rate": 2.454545454545455e-06, |
|
"loss": 0.0001, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 9.227272727272727, |
|
"grad_norm": 0.0014936975203454494, |
|
"learning_rate": 2.318181818181818e-06, |
|
"loss": 0.0001, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 9.272727272727273, |
|
"grad_norm": 0.0020053344778716564, |
|
"learning_rate": 2.181818181818182e-06, |
|
"loss": 0.0, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 9.318181818181818, |
|
"grad_norm": 0.0015693982131779194, |
|
"learning_rate": 2.0454545454545453e-06, |
|
"loss": 0.0, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 9.363636363636363, |
|
"grad_norm": 0.0016571198357269168, |
|
"learning_rate": 1.909090909090909e-06, |
|
"loss": 0.0001, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 9.409090909090908, |
|
"grad_norm": 0.0013354700058698654, |
|
"learning_rate": 1.7727272727272727e-06, |
|
"loss": 0.0, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 9.454545454545455, |
|
"grad_norm": 0.0013917312026023865, |
|
"learning_rate": 1.6363636363636363e-06, |
|
"loss": 0.0, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"grad_norm": 0.0014988429611548781, |
|
"learning_rate": 1.5e-06, |
|
"loss": 0.0, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 9.545454545454545, |
|
"grad_norm": 0.0014679876621812582, |
|
"learning_rate": 1.3636363636363636e-06, |
|
"loss": 0.0, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 9.590909090909092, |
|
"grad_norm": 0.0018639364279806614, |
|
"learning_rate": 1.2272727272727274e-06, |
|
"loss": 0.0001, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 9.636363636363637, |
|
"grad_norm": 0.001416134531609714, |
|
"learning_rate": 1.090909090909091e-06, |
|
"loss": 0.0, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 9.681818181818182, |
|
"grad_norm": 0.0018406022572889924, |
|
"learning_rate": 9.545454545454546e-07, |
|
"loss": 0.0001, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 9.727272727272727, |
|
"grad_norm": 0.0014891604660078883, |
|
"learning_rate": 8.181818181818181e-07, |
|
"loss": 0.0001, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 9.772727272727273, |
|
"grad_norm": 0.0017427564598619938, |
|
"learning_rate": 6.818181818181818e-07, |
|
"loss": 0.0, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 9.818181818181818, |
|
"grad_norm": 0.0014797528274357319, |
|
"learning_rate": 5.454545454545455e-07, |
|
"loss": 0.0, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 9.863636363636363, |
|
"grad_norm": 0.0015343551058322191, |
|
"learning_rate": 4.0909090909090906e-07, |
|
"loss": 0.0001, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 9.909090909090908, |
|
"grad_norm": 0.0014587711775675416, |
|
"learning_rate": 2.7272727272727274e-07, |
|
"loss": 0.0, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 9.954545454545455, |
|
"grad_norm": 0.0013577837962657213, |
|
"learning_rate": 1.3636363636363637e-07, |
|
"loss": 0.0353, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 0.0017256715800613165, |
|
"learning_rate": 0.0, |
|
"loss": 0.0001, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 3.619581912062131e-05, |
|
"eval_runtime": 1.0359, |
|
"eval_samples_per_second": 847.586, |
|
"eval_steps_per_second": 53.095, |
|
"step": 2200 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2200, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 234554255855400.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|