|
{ |
|
"best_global_step": 40000, |
|
"best_metric": 0.198301300406456, |
|
"best_model_checkpoint": "/media/user/Expansion1/multilingual-e5-small-aligned-v2-fineweb2hq-vs-c4-classifier-run2/checkpoint-40000", |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 120000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0125, |
|
"grad_norm": 5.874658584594727, |
|
"learning_rate": 4.979208333333333e-05, |
|
"loss": 0.3436, |
|
"num_input_tokens_seen": 512000, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.025, |
|
"grad_norm": 0.6375065445899963, |
|
"learning_rate": 4.958375e-05, |
|
"loss": 0.281, |
|
"num_input_tokens_seen": 1024000, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.0375, |
|
"grad_norm": 0.48116546869277954, |
|
"learning_rate": 4.937541666666667e-05, |
|
"loss": 0.2595, |
|
"num_input_tokens_seen": 1536000, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.6305193305015564, |
|
"learning_rate": 4.9167083333333336e-05, |
|
"loss": 0.2466, |
|
"num_input_tokens_seen": 2048000, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.0625, |
|
"grad_norm": 3.560100555419922, |
|
"learning_rate": 4.8958750000000006e-05, |
|
"loss": 0.2283, |
|
"num_input_tokens_seen": 2560000, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.075, |
|
"grad_norm": 43.007347106933594, |
|
"learning_rate": 4.875041666666667e-05, |
|
"loss": 0.2106, |
|
"num_input_tokens_seen": 3072000, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.0875, |
|
"grad_norm": 27.75348472595215, |
|
"learning_rate": 4.854208333333334e-05, |
|
"loss": 0.2198, |
|
"num_input_tokens_seen": 3584000, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.11089582741260529, |
|
"learning_rate": 4.833375e-05, |
|
"loss": 0.2145, |
|
"num_input_tokens_seen": 4096000, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.1125, |
|
"grad_norm": 0.08869020640850067, |
|
"learning_rate": 4.812541666666667e-05, |
|
"loss": 0.2033, |
|
"num_input_tokens_seen": 4608000, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.125, |
|
"grad_norm": 22.97330665588379, |
|
"learning_rate": 4.791708333333333e-05, |
|
"loss": 0.2007, |
|
"num_input_tokens_seen": 5120000, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.1375, |
|
"grad_norm": 0.12722039222717285, |
|
"learning_rate": 4.770875e-05, |
|
"loss": 0.1893, |
|
"num_input_tokens_seen": 5632000, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 24.923276901245117, |
|
"learning_rate": 4.750041666666667e-05, |
|
"loss": 0.2073, |
|
"num_input_tokens_seen": 6144000, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.1625, |
|
"grad_norm": 0.17447435855865479, |
|
"learning_rate": 4.729208333333334e-05, |
|
"loss": 0.2024, |
|
"num_input_tokens_seen": 6656000, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.175, |
|
"grad_norm": 5.2057671546936035, |
|
"learning_rate": 4.708375e-05, |
|
"loss": 0.2168, |
|
"num_input_tokens_seen": 7168000, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.1875, |
|
"grad_norm": 0.35154542326927185, |
|
"learning_rate": 4.687541666666667e-05, |
|
"loss": 0.2002, |
|
"num_input_tokens_seen": 7680000, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.17855411767959595, |
|
"learning_rate": 4.6667083333333336e-05, |
|
"loss": 0.1863, |
|
"num_input_tokens_seen": 8192000, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.2125, |
|
"grad_norm": 0.26302626729011536, |
|
"learning_rate": 4.645875e-05, |
|
"loss": 0.2023, |
|
"num_input_tokens_seen": 8704000, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.225, |
|
"grad_norm": 1.0355998277664185, |
|
"learning_rate": 4.625041666666667e-05, |
|
"loss": 0.1996, |
|
"num_input_tokens_seen": 9216000, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.2375, |
|
"grad_norm": 3.5437514781951904, |
|
"learning_rate": 4.6042083333333334e-05, |
|
"loss": 0.1928, |
|
"num_input_tokens_seen": 9728000, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 18.394441604614258, |
|
"learning_rate": 4.5833750000000005e-05, |
|
"loss": 0.1968, |
|
"num_input_tokens_seen": 10240000, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.2625, |
|
"grad_norm": 2.9275922775268555, |
|
"learning_rate": 4.562541666666667e-05, |
|
"loss": 0.2081, |
|
"num_input_tokens_seen": 10752000, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.275, |
|
"grad_norm": 3.922402858734131, |
|
"learning_rate": 4.541708333333334e-05, |
|
"loss": 0.1769, |
|
"num_input_tokens_seen": 11264000, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.2875, |
|
"grad_norm": 0.25219565629959106, |
|
"learning_rate": 4.5208749999999996e-05, |
|
"loss": 0.1722, |
|
"num_input_tokens_seen": 11776000, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.10917416214942932, |
|
"learning_rate": 4.500041666666667e-05, |
|
"loss": 0.1638, |
|
"num_input_tokens_seen": 12288000, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.3125, |
|
"grad_norm": 0.1928466558456421, |
|
"learning_rate": 4.479208333333334e-05, |
|
"loss": 0.1826, |
|
"num_input_tokens_seen": 12800000, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.325, |
|
"grad_norm": 1.1688095331192017, |
|
"learning_rate": 4.458375e-05, |
|
"loss": 0.186, |
|
"num_input_tokens_seen": 13312000, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.3375, |
|
"grad_norm": 0.07879551500082016, |
|
"learning_rate": 4.437541666666667e-05, |
|
"loss": 0.1638, |
|
"num_input_tokens_seen": 13824000, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 1.294481635093689, |
|
"learning_rate": 4.4167083333333336e-05, |
|
"loss": 0.1695, |
|
"num_input_tokens_seen": 14336000, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.3625, |
|
"grad_norm": 21.836345672607422, |
|
"learning_rate": 4.395875000000001e-05, |
|
"loss": 0.1713, |
|
"num_input_tokens_seen": 14848000, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.375, |
|
"grad_norm": 0.1329026073217392, |
|
"learning_rate": 4.375041666666667e-05, |
|
"loss": 0.178, |
|
"num_input_tokens_seen": 15360000, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.3875, |
|
"grad_norm": 2.196962594985962, |
|
"learning_rate": 4.3542083333333334e-05, |
|
"loss": 0.1616, |
|
"num_input_tokens_seen": 15872000, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.06707575172185898, |
|
"learning_rate": 4.333375e-05, |
|
"loss": 0.1563, |
|
"num_input_tokens_seen": 16384000, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.4125, |
|
"grad_norm": 4.03335428237915, |
|
"learning_rate": 4.312541666666667e-05, |
|
"loss": 0.1644, |
|
"num_input_tokens_seen": 16896000, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.425, |
|
"grad_norm": 0.7723323702812195, |
|
"learning_rate": 4.291708333333333e-05, |
|
"loss": 0.1695, |
|
"num_input_tokens_seen": 17408000, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.4375, |
|
"grad_norm": 0.14005227386951447, |
|
"learning_rate": 4.270875e-05, |
|
"loss": 0.1771, |
|
"num_input_tokens_seen": 17920000, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.22879253327846527, |
|
"learning_rate": 4.250041666666667e-05, |
|
"loss": 0.1652, |
|
"num_input_tokens_seen": 18432000, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.4625, |
|
"grad_norm": 41.407352447509766, |
|
"learning_rate": 4.229208333333334e-05, |
|
"loss": 0.1697, |
|
"num_input_tokens_seen": 18944000, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.475, |
|
"grad_norm": 0.2552178204059601, |
|
"learning_rate": 4.208375e-05, |
|
"loss": 0.169, |
|
"num_input_tokens_seen": 19456000, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.4875, |
|
"grad_norm": 4.208723545074463, |
|
"learning_rate": 4.1875416666666666e-05, |
|
"loss": 0.1887, |
|
"num_input_tokens_seen": 19968000, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.6019179821014404, |
|
"learning_rate": 4.1667083333333336e-05, |
|
"loss": 0.149, |
|
"num_input_tokens_seen": 20480000, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.5125, |
|
"grad_norm": 11.652756690979004, |
|
"learning_rate": 4.145875e-05, |
|
"loss": 0.1485, |
|
"num_input_tokens_seen": 20992000, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.525, |
|
"grad_norm": 3.790767192840576, |
|
"learning_rate": 4.125041666666667e-05, |
|
"loss": 0.1613, |
|
"num_input_tokens_seen": 21504000, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.5375, |
|
"grad_norm": 0.0643579512834549, |
|
"learning_rate": 4.1042083333333335e-05, |
|
"loss": 0.156, |
|
"num_input_tokens_seen": 22016000, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 6.347803115844727, |
|
"learning_rate": 4.0833750000000005e-05, |
|
"loss": 0.1743, |
|
"num_input_tokens_seen": 22528000, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.5625, |
|
"grad_norm": 0.45395129919052124, |
|
"learning_rate": 4.062541666666667e-05, |
|
"loss": 0.1566, |
|
"num_input_tokens_seen": 23040000, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.575, |
|
"grad_norm": 4.8587470054626465, |
|
"learning_rate": 4.041708333333333e-05, |
|
"loss": 0.156, |
|
"num_input_tokens_seen": 23552000, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.5875, |
|
"grad_norm": 0.8771836757659912, |
|
"learning_rate": 4.0208750000000004e-05, |
|
"loss": 0.1473, |
|
"num_input_tokens_seen": 24064000, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 7.632213592529297, |
|
"learning_rate": 4.000041666666667e-05, |
|
"loss": 0.1523, |
|
"num_input_tokens_seen": 24576000, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.6125, |
|
"grad_norm": 4.013466835021973, |
|
"learning_rate": 3.979208333333334e-05, |
|
"loss": 0.1646, |
|
"num_input_tokens_seen": 25088000, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.625, |
|
"grad_norm": 0.08471076935529709, |
|
"learning_rate": 3.958375e-05, |
|
"loss": 0.1653, |
|
"num_input_tokens_seen": 25600000, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.6375, |
|
"grad_norm": 4.263392448425293, |
|
"learning_rate": 3.937541666666667e-05, |
|
"loss": 0.1513, |
|
"num_input_tokens_seen": 26112000, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 17.002439498901367, |
|
"learning_rate": 3.9167083333333336e-05, |
|
"loss": 0.144, |
|
"num_input_tokens_seen": 26624000, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.6625, |
|
"grad_norm": 0.056964244693517685, |
|
"learning_rate": 3.895875e-05, |
|
"loss": 0.1517, |
|
"num_input_tokens_seen": 27136000, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.675, |
|
"grad_norm": 0.4403606653213501, |
|
"learning_rate": 3.8750416666666664e-05, |
|
"loss": 0.1501, |
|
"num_input_tokens_seen": 27648000, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.6875, |
|
"grad_norm": 0.11431742459535599, |
|
"learning_rate": 3.8542083333333335e-05, |
|
"loss": 0.1597, |
|
"num_input_tokens_seen": 28160000, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 7.810299396514893, |
|
"learning_rate": 3.833375e-05, |
|
"loss": 0.1437, |
|
"num_input_tokens_seen": 28672000, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.7125, |
|
"grad_norm": 5.501016616821289, |
|
"learning_rate": 3.812541666666667e-05, |
|
"loss": 0.1415, |
|
"num_input_tokens_seen": 29184000, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.725, |
|
"grad_norm": 0.20471607148647308, |
|
"learning_rate": 3.791708333333333e-05, |
|
"loss": 0.1545, |
|
"num_input_tokens_seen": 29696000, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.7375, |
|
"grad_norm": 4.6759748458862305, |
|
"learning_rate": 3.7708750000000004e-05, |
|
"loss": 0.1431, |
|
"num_input_tokens_seen": 30208000, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.12430200725793839, |
|
"learning_rate": 3.7500416666666674e-05, |
|
"loss": 0.1337, |
|
"num_input_tokens_seen": 30720000, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.7625, |
|
"grad_norm": 0.19309231638908386, |
|
"learning_rate": 3.729208333333333e-05, |
|
"loss": 0.1338, |
|
"num_input_tokens_seen": 31232000, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.775, |
|
"grad_norm": 0.18272100389003754, |
|
"learning_rate": 3.708375e-05, |
|
"loss": 0.1504, |
|
"num_input_tokens_seen": 31744000, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.7875, |
|
"grad_norm": 0.08077079057693481, |
|
"learning_rate": 3.6875416666666666e-05, |
|
"loss": 0.13, |
|
"num_input_tokens_seen": 32256000, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.3213728368282318, |
|
"learning_rate": 3.666708333333334e-05, |
|
"loss": 0.1406, |
|
"num_input_tokens_seen": 32768000, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.8125, |
|
"grad_norm": 2.6233737468719482, |
|
"learning_rate": 3.645875e-05, |
|
"loss": 0.1435, |
|
"num_input_tokens_seen": 33280000, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 0.825, |
|
"grad_norm": 0.1083766371011734, |
|
"learning_rate": 3.625041666666667e-05, |
|
"loss": 0.1361, |
|
"num_input_tokens_seen": 33792000, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.8375, |
|
"grad_norm": 0.08282948285341263, |
|
"learning_rate": 3.6042083333333335e-05, |
|
"loss": 0.1483, |
|
"num_input_tokens_seen": 34304000, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.10612857341766357, |
|
"learning_rate": 3.583375e-05, |
|
"loss": 0.1419, |
|
"num_input_tokens_seen": 34816000, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 0.8625, |
|
"grad_norm": 1.134366750717163, |
|
"learning_rate": 3.562541666666667e-05, |
|
"loss": 0.1381, |
|
"num_input_tokens_seen": 35328000, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 0.875, |
|
"grad_norm": 32.71923065185547, |
|
"learning_rate": 3.5417083333333333e-05, |
|
"loss": 0.1395, |
|
"num_input_tokens_seen": 35840000, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 0.8875, |
|
"grad_norm": 0.14211738109588623, |
|
"learning_rate": 3.5208750000000004e-05, |
|
"loss": 0.1543, |
|
"num_input_tokens_seen": 36352000, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.04908756539225578, |
|
"learning_rate": 3.500041666666667e-05, |
|
"loss": 0.1229, |
|
"num_input_tokens_seen": 36864000, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 0.9125, |
|
"grad_norm": 0.47926560044288635, |
|
"learning_rate": 3.479208333333334e-05, |
|
"loss": 0.1251, |
|
"num_input_tokens_seen": 37376000, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 0.925, |
|
"grad_norm": 1.389073371887207, |
|
"learning_rate": 3.458375e-05, |
|
"loss": 0.1718, |
|
"num_input_tokens_seen": 37888000, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 0.9375, |
|
"grad_norm": 0.14962467551231384, |
|
"learning_rate": 3.437541666666667e-05, |
|
"loss": 0.1579, |
|
"num_input_tokens_seen": 38400000, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 3.8970487117767334, |
|
"learning_rate": 3.416708333333333e-05, |
|
"loss": 0.1294, |
|
"num_input_tokens_seen": 38912000, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 0.9625, |
|
"grad_norm": 0.1765402853488922, |
|
"learning_rate": 3.395875e-05, |
|
"loss": 0.1612, |
|
"num_input_tokens_seen": 39424000, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 0.975, |
|
"grad_norm": 0.1407538652420044, |
|
"learning_rate": 3.3750416666666665e-05, |
|
"loss": 0.1336, |
|
"num_input_tokens_seen": 39936000, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 0.9875, |
|
"grad_norm": 8.721837997436523, |
|
"learning_rate": 3.3542083333333335e-05, |
|
"loss": 0.1268, |
|
"num_input_tokens_seen": 40448000, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.15711814165115356, |
|
"learning_rate": 3.333375e-05, |
|
"loss": 0.1387, |
|
"num_input_tokens_seen": 40960000, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.951525, |
|
"eval_combined_score": 1.3493705490187269, |
|
"eval_loss": 0.198301300406456, |
|
"eval_runtime": 39.36, |
|
"eval_samples_per_second": 2032.519, |
|
"eval_steps_per_second": 254.065, |
|
"num_input_tokens_seen": 40960000, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.0125, |
|
"grad_norm": 0.02203565090894699, |
|
"learning_rate": 3.312541666666667e-05, |
|
"loss": 0.1092, |
|
"num_input_tokens_seen": 41472000, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 1.025, |
|
"grad_norm": 328.5529479980469, |
|
"learning_rate": 3.291708333333334e-05, |
|
"loss": 0.0844, |
|
"num_input_tokens_seen": 41984000, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.0375, |
|
"grad_norm": 0.11391662806272507, |
|
"learning_rate": 3.270875e-05, |
|
"loss": 0.1115, |
|
"num_input_tokens_seen": 42496000, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.07919144630432129, |
|
"learning_rate": 3.250041666666667e-05, |
|
"loss": 0.0878, |
|
"num_input_tokens_seen": 43008000, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.0625, |
|
"grad_norm": 0.029733452945947647, |
|
"learning_rate": 3.229208333333333e-05, |
|
"loss": 0.0914, |
|
"num_input_tokens_seen": 43520000, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 1.075, |
|
"grad_norm": 0.277174711227417, |
|
"learning_rate": 3.208375e-05, |
|
"loss": 0.1028, |
|
"num_input_tokens_seen": 44032000, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.0875, |
|
"grad_norm": 0.11592718958854675, |
|
"learning_rate": 3.1875416666666666e-05, |
|
"loss": 0.1032, |
|
"num_input_tokens_seen": 44544000, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.2565874755382538, |
|
"learning_rate": 3.166708333333334e-05, |
|
"loss": 0.0912, |
|
"num_input_tokens_seen": 45056000, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.1125, |
|
"grad_norm": 0.03557795658707619, |
|
"learning_rate": 3.145875e-05, |
|
"loss": 0.0882, |
|
"num_input_tokens_seen": 45568000, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 1.125, |
|
"grad_norm": 4.72824764251709, |
|
"learning_rate": 3.125041666666667e-05, |
|
"loss": 0.0918, |
|
"num_input_tokens_seen": 46080000, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 1.1375, |
|
"grad_norm": 0.10075237601995468, |
|
"learning_rate": 3.1042083333333335e-05, |
|
"loss": 0.1117, |
|
"num_input_tokens_seen": 46592000, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.07608671486377716, |
|
"learning_rate": 3.083375e-05, |
|
"loss": 0.1383, |
|
"num_input_tokens_seen": 47104000, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 1.1625, |
|
"grad_norm": 14.187911033630371, |
|
"learning_rate": 3.062541666666667e-05, |
|
"loss": 0.1062, |
|
"num_input_tokens_seen": 47616000, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 1.175, |
|
"grad_norm": 0.0889461487531662, |
|
"learning_rate": 3.0417083333333334e-05, |
|
"loss": 0.1255, |
|
"num_input_tokens_seen": 48128000, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 1.1875, |
|
"grad_norm": 0.10722101479768753, |
|
"learning_rate": 3.020875e-05, |
|
"loss": 0.0987, |
|
"num_input_tokens_seen": 48640000, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 42.01722717285156, |
|
"learning_rate": 3.000041666666667e-05, |
|
"loss": 0.1047, |
|
"num_input_tokens_seen": 49152000, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 1.2125, |
|
"grad_norm": 0.08817047625780106, |
|
"learning_rate": 2.9792083333333336e-05, |
|
"loss": 0.0931, |
|
"num_input_tokens_seen": 49664000, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 1.225, |
|
"grad_norm": 0.058988332748413086, |
|
"learning_rate": 2.958375e-05, |
|
"loss": 0.1048, |
|
"num_input_tokens_seen": 50176000, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 1.2375, |
|
"grad_norm": 0.025500474497675896, |
|
"learning_rate": 2.9375416666666667e-05, |
|
"loss": 0.0845, |
|
"num_input_tokens_seen": 50688000, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.07621040940284729, |
|
"learning_rate": 2.9167083333333334e-05, |
|
"loss": 0.0893, |
|
"num_input_tokens_seen": 51200000, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 1.2625, |
|
"grad_norm": 0.14059029519557953, |
|
"learning_rate": 2.895875e-05, |
|
"loss": 0.0911, |
|
"num_input_tokens_seen": 51712000, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 1.275, |
|
"grad_norm": 0.04900716617703438, |
|
"learning_rate": 2.875041666666667e-05, |
|
"loss": 0.0984, |
|
"num_input_tokens_seen": 52224000, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 1.2875, |
|
"grad_norm": 0.0568259134888649, |
|
"learning_rate": 2.8542083333333336e-05, |
|
"loss": 0.0995, |
|
"num_input_tokens_seen": 52736000, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.052453652024269104, |
|
"learning_rate": 2.8333750000000003e-05, |
|
"loss": 0.0875, |
|
"num_input_tokens_seen": 53248000, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 1.3125, |
|
"grad_norm": 0.05766982212662697, |
|
"learning_rate": 2.812541666666667e-05, |
|
"loss": 0.0772, |
|
"num_input_tokens_seen": 53760000, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 1.325, |
|
"grad_norm": 0.06809753179550171, |
|
"learning_rate": 2.7917083333333334e-05, |
|
"loss": 0.0849, |
|
"num_input_tokens_seen": 54272000, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 1.3375, |
|
"grad_norm": 0.5180730223655701, |
|
"learning_rate": 2.770875e-05, |
|
"loss": 0.089, |
|
"num_input_tokens_seen": 54784000, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 12.439111709594727, |
|
"learning_rate": 2.750041666666667e-05, |
|
"loss": 0.0927, |
|
"num_input_tokens_seen": 55296000, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 1.3625, |
|
"grad_norm": 0.12473966181278229, |
|
"learning_rate": 2.7292083333333336e-05, |
|
"loss": 0.1109, |
|
"num_input_tokens_seen": 55808000, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 1.375, |
|
"grad_norm": 1.5988309383392334, |
|
"learning_rate": 2.7083750000000003e-05, |
|
"loss": 0.0962, |
|
"num_input_tokens_seen": 56320000, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 1.3875, |
|
"grad_norm": 0.09568974375724792, |
|
"learning_rate": 2.687541666666667e-05, |
|
"loss": 0.1041, |
|
"num_input_tokens_seen": 56832000, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 105.98346710205078, |
|
"learning_rate": 2.6667083333333338e-05, |
|
"loss": 0.0895, |
|
"num_input_tokens_seen": 57344000, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 1.4125, |
|
"grad_norm": 0.04284173250198364, |
|
"learning_rate": 2.6458749999999998e-05, |
|
"loss": 0.09, |
|
"num_input_tokens_seen": 57856000, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 1.425, |
|
"grad_norm": 0.06827585399150848, |
|
"learning_rate": 2.6250416666666665e-05, |
|
"loss": 0.1064, |
|
"num_input_tokens_seen": 58368000, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 1.4375, |
|
"grad_norm": 0.10431079566478729, |
|
"learning_rate": 2.6042083333333333e-05, |
|
"loss": 0.1033, |
|
"num_input_tokens_seen": 58880000, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.08796288073062897, |
|
"learning_rate": 2.583375e-05, |
|
"loss": 0.0781, |
|
"num_input_tokens_seen": 59392000, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 1.4625, |
|
"grad_norm": 1.883025884628296, |
|
"learning_rate": 2.5625416666666667e-05, |
|
"loss": 0.1001, |
|
"num_input_tokens_seen": 59904000, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 1.475, |
|
"grad_norm": 0.2901429533958435, |
|
"learning_rate": 2.5417083333333334e-05, |
|
"loss": 0.0965, |
|
"num_input_tokens_seen": 60416000, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 1.4875, |
|
"grad_norm": 0.05163797363638878, |
|
"learning_rate": 2.520875e-05, |
|
"loss": 0.1064, |
|
"num_input_tokens_seen": 60928000, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.03811231628060341, |
|
"learning_rate": 2.5000416666666672e-05, |
|
"loss": 0.0865, |
|
"num_input_tokens_seen": 61440000, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 1.5125, |
|
"grad_norm": 8.308381080627441, |
|
"learning_rate": 2.4792083333333336e-05, |
|
"loss": 0.093, |
|
"num_input_tokens_seen": 61952000, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 1.525, |
|
"grad_norm": 0.03729177638888359, |
|
"learning_rate": 2.458375e-05, |
|
"loss": 0.077, |
|
"num_input_tokens_seen": 62464000, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 1.5375, |
|
"grad_norm": 5.803088188171387, |
|
"learning_rate": 2.4375416666666667e-05, |
|
"loss": 0.101, |
|
"num_input_tokens_seen": 62976000, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 2.017042636871338, |
|
"learning_rate": 2.4167083333333334e-05, |
|
"loss": 0.0896, |
|
"num_input_tokens_seen": 63488000, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 1.5625, |
|
"grad_norm": 32.310630798339844, |
|
"learning_rate": 2.395875e-05, |
|
"loss": 0.1145, |
|
"num_input_tokens_seen": 64000000, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 1.575, |
|
"grad_norm": 0.37863266468048096, |
|
"learning_rate": 2.3750416666666665e-05, |
|
"loss": 0.1017, |
|
"num_input_tokens_seen": 64512000, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 1.5875, |
|
"grad_norm": 0.05939273163676262, |
|
"learning_rate": 2.3542083333333333e-05, |
|
"loss": 0.0962, |
|
"num_input_tokens_seen": 65024000, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.045398563146591187, |
|
"learning_rate": 2.333375e-05, |
|
"loss": 0.0895, |
|
"num_input_tokens_seen": 65536000, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 1.6125, |
|
"grad_norm": 1.5717942714691162, |
|
"learning_rate": 2.3125416666666667e-05, |
|
"loss": 0.0788, |
|
"num_input_tokens_seen": 66048000, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 1.625, |
|
"grad_norm": 6.278552532196045, |
|
"learning_rate": 2.2917083333333334e-05, |
|
"loss": 0.0876, |
|
"num_input_tokens_seen": 66560000, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 1.6375, |
|
"grad_norm": 0.048090457916259766, |
|
"learning_rate": 2.2708750000000002e-05, |
|
"loss": 0.0709, |
|
"num_input_tokens_seen": 67072000, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.5121225714683533, |
|
"learning_rate": 2.250041666666667e-05, |
|
"loss": 0.0906, |
|
"num_input_tokens_seen": 67584000, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 1.6625, |
|
"grad_norm": 0.04399065673351288, |
|
"learning_rate": 2.2292083333333336e-05, |
|
"loss": 0.097, |
|
"num_input_tokens_seen": 68096000, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 1.675, |
|
"grad_norm": 0.051211412996053696, |
|
"learning_rate": 2.208375e-05, |
|
"loss": 0.0931, |
|
"num_input_tokens_seen": 68608000, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 1.6875, |
|
"grad_norm": 58.19650650024414, |
|
"learning_rate": 2.1875416666666667e-05, |
|
"loss": 0.0794, |
|
"num_input_tokens_seen": 69120000, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.07303386926651001, |
|
"learning_rate": 2.1667083333333335e-05, |
|
"loss": 0.1015, |
|
"num_input_tokens_seen": 69632000, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 1.7125, |
|
"grad_norm": 0.02853875607252121, |
|
"learning_rate": 2.145875e-05, |
|
"loss": 0.0684, |
|
"num_input_tokens_seen": 70144000, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 1.725, |
|
"grad_norm": 0.06810135394334793, |
|
"learning_rate": 2.1250416666666666e-05, |
|
"loss": 0.101, |
|
"num_input_tokens_seen": 70656000, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 1.7375, |
|
"grad_norm": 0.019835174083709717, |
|
"learning_rate": 2.1042083333333333e-05, |
|
"loss": 0.0719, |
|
"num_input_tokens_seen": 71168000, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.07850214093923569, |
|
"learning_rate": 2.083375e-05, |
|
"loss": 0.0808, |
|
"num_input_tokens_seen": 71680000, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 1.7625, |
|
"grad_norm": 0.050091035664081573, |
|
"learning_rate": 2.0625416666666667e-05, |
|
"loss": 0.0835, |
|
"num_input_tokens_seen": 72192000, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 1.775, |
|
"grad_norm": 0.01498348731547594, |
|
"learning_rate": 2.0417083333333335e-05, |
|
"loss": 0.0918, |
|
"num_input_tokens_seen": 72704000, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 1.7875, |
|
"grad_norm": 0.034435465931892395, |
|
"learning_rate": 2.0208750000000002e-05, |
|
"loss": 0.0777, |
|
"num_input_tokens_seen": 73216000, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.045340586453676224, |
|
"learning_rate": 2.000041666666667e-05, |
|
"loss": 0.0894, |
|
"num_input_tokens_seen": 73728000, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 1.8125, |
|
"grad_norm": 0.28705134987831116, |
|
"learning_rate": 1.9792083333333333e-05, |
|
"loss": 0.0741, |
|
"num_input_tokens_seen": 74240000, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 1.825, |
|
"grad_norm": 155.12445068359375, |
|
"learning_rate": 1.958375e-05, |
|
"loss": 0.0813, |
|
"num_input_tokens_seen": 74752000, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 1.8375, |
|
"grad_norm": 0.08065774291753769, |
|
"learning_rate": 1.9375416666666668e-05, |
|
"loss": 0.0819, |
|
"num_input_tokens_seen": 75264000, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.40842482447624207, |
|
"learning_rate": 1.9167083333333335e-05, |
|
"loss": 0.0669, |
|
"num_input_tokens_seen": 75776000, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 1.8625, |
|
"grad_norm": 0.09376771003007889, |
|
"learning_rate": 1.895875e-05, |
|
"loss": 0.0749, |
|
"num_input_tokens_seen": 76288000, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 1.875, |
|
"grad_norm": 0.035654786974191666, |
|
"learning_rate": 1.8750416666666666e-05, |
|
"loss": 0.0757, |
|
"num_input_tokens_seen": 76800000, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 1.8875, |
|
"grad_norm": 0.04763418436050415, |
|
"learning_rate": 1.8542083333333337e-05, |
|
"loss": 0.0652, |
|
"num_input_tokens_seen": 77312000, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.04799911379814148, |
|
"learning_rate": 1.833375e-05, |
|
"loss": 0.0688, |
|
"num_input_tokens_seen": 77824000, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 1.9125, |
|
"grad_norm": 0.22011104226112366, |
|
"learning_rate": 1.8125416666666668e-05, |
|
"loss": 0.0797, |
|
"num_input_tokens_seen": 78336000, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 1.925, |
|
"grad_norm": 0.05009845644235611, |
|
"learning_rate": 1.7917083333333335e-05, |
|
"loss": 0.0542, |
|
"num_input_tokens_seen": 78848000, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 1.9375, |
|
"grad_norm": 1.8440918922424316, |
|
"learning_rate": 1.7708750000000002e-05, |
|
"loss": 0.0773, |
|
"num_input_tokens_seen": 79360000, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.025103362277150154, |
|
"learning_rate": 1.7500416666666666e-05, |
|
"loss": 0.0692, |
|
"num_input_tokens_seen": 79872000, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 1.9625, |
|
"grad_norm": 14.473067283630371, |
|
"learning_rate": 1.7292083333333333e-05, |
|
"loss": 0.0658, |
|
"num_input_tokens_seen": 80384000, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 1.975, |
|
"grad_norm": 0.051201172173023224, |
|
"learning_rate": 1.708375e-05, |
|
"loss": 0.0648, |
|
"num_input_tokens_seen": 80896000, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 1.9875, |
|
"grad_norm": 0.05888315662741661, |
|
"learning_rate": 1.6875416666666668e-05, |
|
"loss": 0.0792, |
|
"num_input_tokens_seen": 81408000, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.08364333212375641, |
|
"learning_rate": 1.666708333333333e-05, |
|
"loss": 0.0682, |
|
"num_input_tokens_seen": 81920000, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9528, |
|
"eval_combined_score": 1.327030219983165, |
|
"eval_loss": 0.22639968991279602, |
|
"eval_runtime": 39.5018, |
|
"eval_samples_per_second": 2025.225, |
|
"eval_steps_per_second": 253.153, |
|
"num_input_tokens_seen": 81920000, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 2.0125, |
|
"grad_norm": 1.8050274848937988, |
|
"learning_rate": 1.645875e-05, |
|
"loss": 0.0508, |
|
"num_input_tokens_seen": 82432000, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 2.025, |
|
"grad_norm": 0.02845793031156063, |
|
"learning_rate": 1.625041666666667e-05, |
|
"loss": 0.0525, |
|
"num_input_tokens_seen": 82944000, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 2.0375, |
|
"grad_norm": 0.025743141770362854, |
|
"learning_rate": 1.6042083333333337e-05, |
|
"loss": 0.0483, |
|
"num_input_tokens_seen": 83456000, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 8.399374008178711, |
|
"learning_rate": 1.583375e-05, |
|
"loss": 0.0652, |
|
"num_input_tokens_seen": 83968000, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 2.0625, |
|
"grad_norm": 0.017835861071944237, |
|
"learning_rate": 1.5625416666666668e-05, |
|
"loss": 0.0482, |
|
"num_input_tokens_seen": 84480000, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 2.075, |
|
"grad_norm": 0.06692575663328171, |
|
"learning_rate": 1.5417083333333335e-05, |
|
"loss": 0.0521, |
|
"num_input_tokens_seen": 84992000, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 2.0875, |
|
"grad_norm": 0.035685233771800995, |
|
"learning_rate": 1.5208749999999999e-05, |
|
"loss": 0.0582, |
|
"num_input_tokens_seen": 85504000, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 0.13481773436069489, |
|
"learning_rate": 1.5000416666666666e-05, |
|
"loss": 0.0468, |
|
"num_input_tokens_seen": 86016000, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 2.1125, |
|
"grad_norm": 17.486289978027344, |
|
"learning_rate": 1.4792083333333334e-05, |
|
"loss": 0.0559, |
|
"num_input_tokens_seen": 86528000, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 2.125, |
|
"grad_norm": 0.03201691806316376, |
|
"learning_rate": 1.458375e-05, |
|
"loss": 0.0425, |
|
"num_input_tokens_seen": 87040000, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 2.1375, |
|
"grad_norm": 0.021227147430181503, |
|
"learning_rate": 1.4375416666666666e-05, |
|
"loss": 0.05, |
|
"num_input_tokens_seen": 87552000, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 0.040053412318229675, |
|
"learning_rate": 1.4167083333333334e-05, |
|
"loss": 0.048, |
|
"num_input_tokens_seen": 88064000, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 2.1625, |
|
"grad_norm": 0.01835712045431137, |
|
"learning_rate": 1.395875e-05, |
|
"loss": 0.038, |
|
"num_input_tokens_seen": 88576000, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 2.175, |
|
"grad_norm": 0.03529110550880432, |
|
"learning_rate": 1.3750416666666668e-05, |
|
"loss": 0.051, |
|
"num_input_tokens_seen": 89088000, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 2.1875, |
|
"grad_norm": 97.47898864746094, |
|
"learning_rate": 1.3542083333333334e-05, |
|
"loss": 0.0373, |
|
"num_input_tokens_seen": 89600000, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.014346601441502571, |
|
"learning_rate": 1.3333750000000001e-05, |
|
"loss": 0.0557, |
|
"num_input_tokens_seen": 90112000, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 2.2125, |
|
"grad_norm": 0.01878521591424942, |
|
"learning_rate": 1.3125416666666668e-05, |
|
"loss": 0.0526, |
|
"num_input_tokens_seen": 90624000, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 2.225, |
|
"grad_norm": 0.04132438451051712, |
|
"learning_rate": 1.2917083333333335e-05, |
|
"loss": 0.0464, |
|
"num_input_tokens_seen": 91136000, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 2.2375, |
|
"grad_norm": 0.02642699144780636, |
|
"learning_rate": 1.270875e-05, |
|
"loss": 0.0464, |
|
"num_input_tokens_seen": 91648000, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 877.631591796875, |
|
"learning_rate": 1.2500416666666666e-05, |
|
"loss": 0.0374, |
|
"num_input_tokens_seen": 92160000, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 2.2625, |
|
"grad_norm": 0.4528743326663971, |
|
"learning_rate": 1.2292083333333334e-05, |
|
"loss": 0.059, |
|
"num_input_tokens_seen": 92672000, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 2.275, |
|
"grad_norm": 0.1183973178267479, |
|
"learning_rate": 1.2083750000000001e-05, |
|
"loss": 0.0505, |
|
"num_input_tokens_seen": 93184000, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 2.2875, |
|
"grad_norm": 0.04196188971400261, |
|
"learning_rate": 1.1875416666666667e-05, |
|
"loss": 0.0517, |
|
"num_input_tokens_seen": 93696000, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 0.03194092586636543, |
|
"learning_rate": 1.1667083333333334e-05, |
|
"loss": 0.0511, |
|
"num_input_tokens_seen": 94208000, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 2.3125, |
|
"grad_norm": 0.0184203889220953, |
|
"learning_rate": 1.145875e-05, |
|
"loss": 0.0501, |
|
"num_input_tokens_seen": 94720000, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 2.325, |
|
"grad_norm": 0.053758785128593445, |
|
"learning_rate": 1.1250416666666667e-05, |
|
"loss": 0.0412, |
|
"num_input_tokens_seen": 95232000, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 2.3375, |
|
"grad_norm": 0.036872394382953644, |
|
"learning_rate": 1.1042083333333334e-05, |
|
"loss": 0.0418, |
|
"num_input_tokens_seen": 95744000, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 0.014922083355486393, |
|
"learning_rate": 1.0833750000000001e-05, |
|
"loss": 0.0507, |
|
"num_input_tokens_seen": 96256000, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 2.3625, |
|
"grad_norm": 0.020171664655208588, |
|
"learning_rate": 1.0625416666666667e-05, |
|
"loss": 0.0444, |
|
"num_input_tokens_seen": 96768000, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 2.375, |
|
"grad_norm": 1.7357385158538818, |
|
"learning_rate": 1.0417083333333334e-05, |
|
"loss": 0.0545, |
|
"num_input_tokens_seen": 97280000, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 2.3875, |
|
"grad_norm": 0.023304857313632965, |
|
"learning_rate": 1.020875e-05, |
|
"loss": 0.0515, |
|
"num_input_tokens_seen": 97792000, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 0.01939631998538971, |
|
"learning_rate": 1.0000416666666668e-05, |
|
"loss": 0.0495, |
|
"num_input_tokens_seen": 98304000, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 2.4125, |
|
"grad_norm": 0.019845569506287575, |
|
"learning_rate": 9.792083333333334e-06, |
|
"loss": 0.0411, |
|
"num_input_tokens_seen": 98816000, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 2.425, |
|
"grad_norm": 0.014959324151277542, |
|
"learning_rate": 9.583750000000001e-06, |
|
"loss": 0.0342, |
|
"num_input_tokens_seen": 99328000, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 2.4375, |
|
"grad_norm": 0.01649474911391735, |
|
"learning_rate": 9.375416666666667e-06, |
|
"loss": 0.0472, |
|
"num_input_tokens_seen": 99840000, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 0.031929146498441696, |
|
"learning_rate": 9.167083333333332e-06, |
|
"loss": 0.0384, |
|
"num_input_tokens_seen": 100352000, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 2.4625, |
|
"grad_norm": 0.10980285704135895, |
|
"learning_rate": 8.958750000000001e-06, |
|
"loss": 0.0513, |
|
"num_input_tokens_seen": 100864000, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 2.475, |
|
"grad_norm": 0.020933426916599274, |
|
"learning_rate": 8.750416666666667e-06, |
|
"loss": 0.0478, |
|
"num_input_tokens_seen": 101376000, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 2.4875, |
|
"grad_norm": 0.017404716461896896, |
|
"learning_rate": 8.542083333333334e-06, |
|
"loss": 0.039, |
|
"num_input_tokens_seen": 101888000, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.11211936920881271, |
|
"learning_rate": 8.33375e-06, |
|
"loss": 0.0473, |
|
"num_input_tokens_seen": 102400000, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 2.5125, |
|
"grad_norm": 0.023433908820152283, |
|
"learning_rate": 8.125416666666667e-06, |
|
"loss": 0.0417, |
|
"num_input_tokens_seen": 102912000, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 2.525, |
|
"grad_norm": 0.011504637077450752, |
|
"learning_rate": 7.917083333333334e-06, |
|
"loss": 0.0396, |
|
"num_input_tokens_seen": 103424000, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 2.5375, |
|
"grad_norm": 0.012843768112361431, |
|
"learning_rate": 7.708750000000001e-06, |
|
"loss": 0.0475, |
|
"num_input_tokens_seen": 103936000, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 0.020524220541119576, |
|
"learning_rate": 7.500416666666667e-06, |
|
"loss": 0.0472, |
|
"num_input_tokens_seen": 104448000, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 2.5625, |
|
"grad_norm": 0.01861303672194481, |
|
"learning_rate": 7.292083333333334e-06, |
|
"loss": 0.0294, |
|
"num_input_tokens_seen": 104960000, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 2.575, |
|
"grad_norm": 0.021343663334846497, |
|
"learning_rate": 7.08375e-06, |
|
"loss": 0.0376, |
|
"num_input_tokens_seen": 105472000, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 2.5875, |
|
"grad_norm": 0.8902124166488647, |
|
"learning_rate": 6.875416666666668e-06, |
|
"loss": 0.0456, |
|
"num_input_tokens_seen": 105984000, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 0.06912536174058914, |
|
"learning_rate": 6.667083333333333e-06, |
|
"loss": 0.0365, |
|
"num_input_tokens_seen": 106496000, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 2.6125, |
|
"grad_norm": 0.17968295514583588, |
|
"learning_rate": 6.458750000000001e-06, |
|
"loss": 0.0372, |
|
"num_input_tokens_seen": 107008000, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 2.625, |
|
"grad_norm": 0.01591988280415535, |
|
"learning_rate": 6.250416666666667e-06, |
|
"loss": 0.0334, |
|
"num_input_tokens_seen": 107520000, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 2.6375, |
|
"grad_norm": 0.015928415581583977, |
|
"learning_rate": 6.0420833333333334e-06, |
|
"loss": 0.0425, |
|
"num_input_tokens_seen": 108032000, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 0.04025963693857193, |
|
"learning_rate": 5.833750000000001e-06, |
|
"loss": 0.0371, |
|
"num_input_tokens_seen": 108544000, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 2.6625, |
|
"grad_norm": 0.07384547591209412, |
|
"learning_rate": 5.625416666666667e-06, |
|
"loss": 0.043, |
|
"num_input_tokens_seen": 109056000, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 2.675, |
|
"grad_norm": 0.0424518883228302, |
|
"learning_rate": 5.4170833333333335e-06, |
|
"loss": 0.0349, |
|
"num_input_tokens_seen": 109568000, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 2.6875, |
|
"grad_norm": 0.01949002780020237, |
|
"learning_rate": 5.208750000000001e-06, |
|
"loss": 0.04, |
|
"num_input_tokens_seen": 110080000, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 0.011860487051308155, |
|
"learning_rate": 5.000416666666667e-06, |
|
"loss": 0.0274, |
|
"num_input_tokens_seen": 110592000, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 2.7125, |
|
"grad_norm": 0.010354108177125454, |
|
"learning_rate": 4.7920833333333335e-06, |
|
"loss": 0.0305, |
|
"num_input_tokens_seen": 111104000, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 2.725, |
|
"grad_norm": 13.753798484802246, |
|
"learning_rate": 4.583750000000001e-06, |
|
"loss": 0.0446, |
|
"num_input_tokens_seen": 111616000, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 2.7375, |
|
"grad_norm": 0.017097918316721916, |
|
"learning_rate": 4.375416666666666e-06, |
|
"loss": 0.0398, |
|
"num_input_tokens_seen": 112128000, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 0.01976764015853405, |
|
"learning_rate": 4.167083333333334e-06, |
|
"loss": 0.0318, |
|
"num_input_tokens_seen": 112640000, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 2.7625, |
|
"grad_norm": 6.2214531898498535, |
|
"learning_rate": 3.95875e-06, |
|
"loss": 0.0424, |
|
"num_input_tokens_seen": 113152000, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 2.775, |
|
"grad_norm": 0.007995002903044224, |
|
"learning_rate": 3.750416666666667e-06, |
|
"loss": 0.0442, |
|
"num_input_tokens_seen": 113664000, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 2.7875, |
|
"grad_norm": 0.05575885996222496, |
|
"learning_rate": 3.5420833333333332e-06, |
|
"loss": 0.037, |
|
"num_input_tokens_seen": 114176000, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 0.1336035579442978, |
|
"learning_rate": 3.33375e-06, |
|
"loss": 0.0385, |
|
"num_input_tokens_seen": 114688000, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 2.8125, |
|
"grad_norm": 0.019037237390875816, |
|
"learning_rate": 3.125416666666667e-06, |
|
"loss": 0.0326, |
|
"num_input_tokens_seen": 115200000, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 2.825, |
|
"grad_norm": 0.00875825248658657, |
|
"learning_rate": 2.9170833333333333e-06, |
|
"loss": 0.0277, |
|
"num_input_tokens_seen": 115712000, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 2.8375, |
|
"grad_norm": 0.052571795880794525, |
|
"learning_rate": 2.70875e-06, |
|
"loss": 0.037, |
|
"num_input_tokens_seen": 116224000, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 0.019845254719257355, |
|
"learning_rate": 2.500416666666667e-06, |
|
"loss": 0.0445, |
|
"num_input_tokens_seen": 116736000, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 2.8625, |
|
"grad_norm": 0.01046211551874876, |
|
"learning_rate": 2.2920833333333338e-06, |
|
"loss": 0.0294, |
|
"num_input_tokens_seen": 117248000, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 2.875, |
|
"grad_norm": 0.012272953987121582, |
|
"learning_rate": 2.0837499999999997e-06, |
|
"loss": 0.0396, |
|
"num_input_tokens_seen": 117760000, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 2.8875, |
|
"grad_norm": 0.022472262382507324, |
|
"learning_rate": 1.8754166666666666e-06, |
|
"loss": 0.047, |
|
"num_input_tokens_seen": 118272000, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 0.07120255380868912, |
|
"learning_rate": 1.6670833333333334e-06, |
|
"loss": 0.0437, |
|
"num_input_tokens_seen": 118784000, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 2.9125, |
|
"grad_norm": 0.060723673552274704, |
|
"learning_rate": 1.45875e-06, |
|
"loss": 0.035, |
|
"num_input_tokens_seen": 119296000, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 2.925, |
|
"grad_norm": 0.012350406497716904, |
|
"learning_rate": 1.2504166666666668e-06, |
|
"loss": 0.03, |
|
"num_input_tokens_seen": 119808000, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 2.9375, |
|
"grad_norm": 0.18025244772434235, |
|
"learning_rate": 1.0420833333333334e-06, |
|
"loss": 0.0306, |
|
"num_input_tokens_seen": 120320000, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 0.007316856179386377, |
|
"learning_rate": 8.3375e-07, |
|
"loss": 0.0328, |
|
"num_input_tokens_seen": 120832000, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 2.9625, |
|
"grad_norm": 0.06496240198612213, |
|
"learning_rate": 6.254166666666667e-07, |
|
"loss": 0.039, |
|
"num_input_tokens_seen": 121344000, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 2.975, |
|
"grad_norm": 0.0057182470336556435, |
|
"learning_rate": 4.170833333333334e-07, |
|
"loss": 0.0367, |
|
"num_input_tokens_seen": 121856000, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 2.9875, |
|
"grad_norm": 0.017466630786657333, |
|
"learning_rate": 2.0875e-07, |
|
"loss": 0.0381, |
|
"num_input_tokens_seen": 122368000, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.013719202019274235, |
|
"learning_rate": 4.1666666666666673e-10, |
|
"loss": 0.0424, |
|
"num_input_tokens_seen": 122880000, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.955225, |
|
"eval_combined_score": 1.2845397902488602, |
|
"eval_loss": 0.2597787082195282, |
|
"eval_runtime": 39.5285, |
|
"eval_samples_per_second": 2023.856, |
|
"eval_steps_per_second": 252.982, |
|
"num_input_tokens_seen": 122880000, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"num_input_tokens_seen": 122880000, |
|
"step": 120000, |
|
"total_flos": 1.580945522688e+16, |
|
"train_loss": 0.10143799341519674, |
|
"train_runtime": 4258.2582, |
|
"train_samples_per_second": 225.444, |
|
"train_steps_per_second": 28.181, |
|
"train_tokens_per_second": 28856.869 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 120000, |
|
"num_input_tokens_seen": 122880000, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.580945522688e+16, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|