|
{
|
|
"best_global_step": 12292,
|
|
"best_metric": 0.9563834240267894,
|
|
"best_model_checkpoint": "C:\\Users\\Shara\\projects\\models\\run_16-lr_2e-05-acc_1-wd_0.01-bs_8-ep_7\\checkpoint-12292",
|
|
"epoch": 7.0,
|
|
"eval_steps": 500,
|
|
"global_step": 12292,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.02847380410022779,
|
|
"grad_norm": 4.8302321434021,
|
|
"learning_rate": 1.992840872111943e-05,
|
|
"loss": 0.9994,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 0.05694760820045558,
|
|
"grad_norm": 3.4978649616241455,
|
|
"learning_rate": 1.9847054995118778e-05,
|
|
"loss": 0.3939,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.08542141230068337,
|
|
"grad_norm": 8.353672981262207,
|
|
"learning_rate": 1.9765701269118128e-05,
|
|
"loss": 0.2882,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 0.11389521640091116,
|
|
"grad_norm": 9.822871208190918,
|
|
"learning_rate": 1.9684347543117477e-05,
|
|
"loss": 0.1728,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.14236902050113895,
|
|
"grad_norm": 3.1411221027374268,
|
|
"learning_rate": 1.9602993817116827e-05,
|
|
"loss": 0.1251,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 0.17084282460136674,
|
|
"grad_norm": 8.493424415588379,
|
|
"learning_rate": 1.9521640091116173e-05,
|
|
"loss": 0.1405,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 0.19931662870159453,
|
|
"grad_norm": 4.757007122039795,
|
|
"learning_rate": 1.9440286365115523e-05,
|
|
"loss": 0.1431,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 0.22779043280182232,
|
|
"grad_norm": 4.339049816131592,
|
|
"learning_rate": 1.9358932639114873e-05,
|
|
"loss": 0.0912,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 0.25626423690205014,
|
|
"grad_norm": 14.540772438049316,
|
|
"learning_rate": 1.9277578913114222e-05,
|
|
"loss": 0.1048,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 0.2847380410022779,
|
|
"grad_norm": 10.503073692321777,
|
|
"learning_rate": 1.9196225187113572e-05,
|
|
"loss": 0.1224,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 0.3132118451025057,
|
|
"grad_norm": 2.738131284713745,
|
|
"learning_rate": 1.911487146111292e-05,
|
|
"loss": 0.1025,
|
|
"step": 550
|
|
},
|
|
{
|
|
"epoch": 0.3416856492027335,
|
|
"grad_norm": 6.681102752685547,
|
|
"learning_rate": 1.903351773511227e-05,
|
|
"loss": 0.0869,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 0.3701594533029613,
|
|
"grad_norm": 1.4330339431762695,
|
|
"learning_rate": 1.895216400911162e-05,
|
|
"loss": 0.052,
|
|
"step": 650
|
|
},
|
|
{
|
|
"epoch": 0.39863325740318906,
|
|
"grad_norm": 2.581470012664795,
|
|
"learning_rate": 1.8870810283110967e-05,
|
|
"loss": 0.0808,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 0.4271070615034169,
|
|
"grad_norm": 5.825100898742676,
|
|
"learning_rate": 1.8789456557110317e-05,
|
|
"loss": 0.0716,
|
|
"step": 750
|
|
},
|
|
{
|
|
"epoch": 0.45558086560364464,
|
|
"grad_norm": 1.74166738986969,
|
|
"learning_rate": 1.8708102831109666e-05,
|
|
"loss": 0.0806,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 0.48405466970387245,
|
|
"grad_norm": 7.3097405433654785,
|
|
"learning_rate": 1.8626749105109016e-05,
|
|
"loss": 0.0666,
|
|
"step": 850
|
|
},
|
|
{
|
|
"epoch": 0.5125284738041003,
|
|
"grad_norm": 2.220766067504883,
|
|
"learning_rate": 1.8545395379108362e-05,
|
|
"loss": 0.0768,
|
|
"step": 900
|
|
},
|
|
{
|
|
"epoch": 0.541002277904328,
|
|
"grad_norm": 4.0566511154174805,
|
|
"learning_rate": 1.8464041653107715e-05,
|
|
"loss": 0.0707,
|
|
"step": 950
|
|
},
|
|
{
|
|
"epoch": 0.5694760820045558,
|
|
"grad_norm": 0.2833240330219269,
|
|
"learning_rate": 1.8382687927107065e-05,
|
|
"loss": 0.0858,
|
|
"step": 1000
|
|
},
|
|
{
|
|
"epoch": 0.5979498861047836,
|
|
"grad_norm": 21.131481170654297,
|
|
"learning_rate": 1.8301334201106415e-05,
|
|
"loss": 0.0595,
|
|
"step": 1050
|
|
},
|
|
{
|
|
"epoch": 0.6264236902050114,
|
|
"grad_norm": 27.655433654785156,
|
|
"learning_rate": 1.821998047510576e-05,
|
|
"loss": 0.0466,
|
|
"step": 1100
|
|
},
|
|
{
|
|
"epoch": 0.6548974943052391,
|
|
"grad_norm": 3.4721717834472656,
|
|
"learning_rate": 1.813862674910511e-05,
|
|
"loss": 0.0776,
|
|
"step": 1150
|
|
},
|
|
{
|
|
"epoch": 0.683371298405467,
|
|
"grad_norm": 2.4811441898345947,
|
|
"learning_rate": 1.805727302310446e-05,
|
|
"loss": 0.0647,
|
|
"step": 1200
|
|
},
|
|
{
|
|
"epoch": 0.7118451025056948,
|
|
"grad_norm": 6.169066429138184,
|
|
"learning_rate": 1.797591929710381e-05,
|
|
"loss": 0.0724,
|
|
"step": 1250
|
|
},
|
|
{
|
|
"epoch": 0.7403189066059226,
|
|
"grad_norm": 8.984748840332031,
|
|
"learning_rate": 1.7894565571103156e-05,
|
|
"loss": 0.0838,
|
|
"step": 1300
|
|
},
|
|
{
|
|
"epoch": 0.7687927107061503,
|
|
"grad_norm": 5.944505214691162,
|
|
"learning_rate": 1.7813211845102506e-05,
|
|
"loss": 0.0683,
|
|
"step": 1350
|
|
},
|
|
{
|
|
"epoch": 0.7972665148063781,
|
|
"grad_norm": 0.28812381625175476,
|
|
"learning_rate": 1.7731858119101856e-05,
|
|
"loss": 0.0614,
|
|
"step": 1400
|
|
},
|
|
{
|
|
"epoch": 0.8257403189066059,
|
|
"grad_norm": 6.176011085510254,
|
|
"learning_rate": 1.7650504393101205e-05,
|
|
"loss": 0.0615,
|
|
"step": 1450
|
|
},
|
|
{
|
|
"epoch": 0.8542141230068337,
|
|
"grad_norm": 2.5244405269622803,
|
|
"learning_rate": 1.7569150667100555e-05,
|
|
"loss": 0.0714,
|
|
"step": 1500
|
|
},
|
|
{
|
|
"epoch": 0.8826879271070615,
|
|
"grad_norm": 3.3508074283599854,
|
|
"learning_rate": 1.7487796941099904e-05,
|
|
"loss": 0.0563,
|
|
"step": 1550
|
|
},
|
|
{
|
|
"epoch": 0.9111617312072893,
|
|
"grad_norm": 9.980842590332031,
|
|
"learning_rate": 1.7406443215099254e-05,
|
|
"loss": 0.0647,
|
|
"step": 1600
|
|
},
|
|
{
|
|
"epoch": 0.9396355353075171,
|
|
"grad_norm": 3.6959774494171143,
|
|
"learning_rate": 1.7325089489098604e-05,
|
|
"loss": 0.0682,
|
|
"step": 1650
|
|
},
|
|
{
|
|
"epoch": 0.9681093394077449,
|
|
"grad_norm": 0.626620888710022,
|
|
"learning_rate": 1.724373576309795e-05,
|
|
"loss": 0.0499,
|
|
"step": 1700
|
|
},
|
|
{
|
|
"epoch": 0.9965831435079726,
|
|
"grad_norm": 7.145672798156738,
|
|
"learning_rate": 1.71623820370973e-05,
|
|
"loss": 0.0569,
|
|
"step": 1750
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"eval_f1": 0.9242246437552389,
|
|
"eval_loss": 0.05600914731621742,
|
|
"eval_precision": 0.9206746826987308,
|
|
"eval_recall": 0.927802086839448,
|
|
"eval_runtime": 3.7538,
|
|
"eval_samples_per_second": 865.798,
|
|
"eval_steps_per_second": 108.425,
|
|
"step": 1756
|
|
},
|
|
{
|
|
"epoch": 1.0250569476082005,
|
|
"grad_norm": 4.280463695526123,
|
|
"learning_rate": 1.708102831109665e-05,
|
|
"loss": 0.0673,
|
|
"step": 1800
|
|
},
|
|
{
|
|
"epoch": 1.0535307517084282,
|
|
"grad_norm": 0.06692535430192947,
|
|
"learning_rate": 1.6999674585096e-05,
|
|
"loss": 0.0403,
|
|
"step": 1850
|
|
},
|
|
{
|
|
"epoch": 1.082004555808656,
|
|
"grad_norm": 0.03789375349879265,
|
|
"learning_rate": 1.691832085909535e-05,
|
|
"loss": 0.0496,
|
|
"step": 1900
|
|
},
|
|
{
|
|
"epoch": 1.1104783599088839,
|
|
"grad_norm": 3.698348045349121,
|
|
"learning_rate": 1.6836967133094698e-05,
|
|
"loss": 0.0381,
|
|
"step": 1950
|
|
},
|
|
{
|
|
"epoch": 1.1389521640091116,
|
|
"grad_norm": 2.753032684326172,
|
|
"learning_rate": 1.6755613407094048e-05,
|
|
"loss": 0.0432,
|
|
"step": 2000
|
|
},
|
|
{
|
|
"epoch": 1.1674259681093395,
|
|
"grad_norm": 0.06336130946874619,
|
|
"learning_rate": 1.6674259681093398e-05,
|
|
"loss": 0.0482,
|
|
"step": 2050
|
|
},
|
|
{
|
|
"epoch": 1.1958997722095672,
|
|
"grad_norm": 3.9503254890441895,
|
|
"learning_rate": 1.6592905955092744e-05,
|
|
"loss": 0.0436,
|
|
"step": 2100
|
|
},
|
|
{
|
|
"epoch": 1.224373576309795,
|
|
"grad_norm": 0.08679741621017456,
|
|
"learning_rate": 1.6511552229092093e-05,
|
|
"loss": 0.0346,
|
|
"step": 2150
|
|
},
|
|
{
|
|
"epoch": 1.2528473804100229,
|
|
"grad_norm": 2.1995127201080322,
|
|
"learning_rate": 1.6430198503091443e-05,
|
|
"loss": 0.0312,
|
|
"step": 2200
|
|
},
|
|
{
|
|
"epoch": 1.2813211845102506,
|
|
"grad_norm": 5.7709808349609375,
|
|
"learning_rate": 1.634884477709079e-05,
|
|
"loss": 0.032,
|
|
"step": 2250
|
|
},
|
|
{
|
|
"epoch": 1.3097949886104785,
|
|
"grad_norm": 0.2031829059123993,
|
|
"learning_rate": 1.626749105109014e-05,
|
|
"loss": 0.053,
|
|
"step": 2300
|
|
},
|
|
{
|
|
"epoch": 1.3382687927107062,
|
|
"grad_norm": 4.292157173156738,
|
|
"learning_rate": 1.6186137325089492e-05,
|
|
"loss": 0.0413,
|
|
"step": 2350
|
|
},
|
|
{
|
|
"epoch": 1.366742596810934,
|
|
"grad_norm": 0.07821401953697205,
|
|
"learning_rate": 1.6104783599088842e-05,
|
|
"loss": 0.0304,
|
|
"step": 2400
|
|
},
|
|
{
|
|
"epoch": 1.3952164009111616,
|
|
"grad_norm": 0.5930206179618835,
|
|
"learning_rate": 1.6023429873088188e-05,
|
|
"loss": 0.0525,
|
|
"step": 2450
|
|
},
|
|
{
|
|
"epoch": 1.4236902050113895,
|
|
"grad_norm": 1.6114712953567505,
|
|
"learning_rate": 1.5942076147087538e-05,
|
|
"loss": 0.0391,
|
|
"step": 2500
|
|
},
|
|
{
|
|
"epoch": 1.4521640091116172,
|
|
"grad_norm": 0.28632065653800964,
|
|
"learning_rate": 1.5860722421086887e-05,
|
|
"loss": 0.0491,
|
|
"step": 2550
|
|
},
|
|
{
|
|
"epoch": 1.4806378132118452,
|
|
"grad_norm": 1.0017669200897217,
|
|
"learning_rate": 1.5779368695086237e-05,
|
|
"loss": 0.0395,
|
|
"step": 2600
|
|
},
|
|
{
|
|
"epoch": 1.5091116173120729,
|
|
"grad_norm": 0.03409272059798241,
|
|
"learning_rate": 1.5698014969085583e-05,
|
|
"loss": 0.0399,
|
|
"step": 2650
|
|
},
|
|
{
|
|
"epoch": 1.5375854214123006,
|
|
"grad_norm": 3.5108702182769775,
|
|
"learning_rate": 1.5616661243084933e-05,
|
|
"loss": 0.0424,
|
|
"step": 2700
|
|
},
|
|
{
|
|
"epoch": 1.5660592255125285,
|
|
"grad_norm": 3.5875349044799805,
|
|
"learning_rate": 1.5535307517084283e-05,
|
|
"loss": 0.0497,
|
|
"step": 2750
|
|
},
|
|
{
|
|
"epoch": 1.5945330296127562,
|
|
"grad_norm": 0.10498251020908356,
|
|
"learning_rate": 1.5453953791083632e-05,
|
|
"loss": 0.0461,
|
|
"step": 2800
|
|
},
|
|
{
|
|
"epoch": 1.6230068337129842,
|
|
"grad_norm": 7.618963718414307,
|
|
"learning_rate": 1.5374227139602996e-05,
|
|
"loss": 0.0394,
|
|
"step": 2850
|
|
},
|
|
{
|
|
"epoch": 1.6514806378132119,
|
|
"grad_norm": 3.3920793533325195,
|
|
"learning_rate": 1.5292873413602346e-05,
|
|
"loss": 0.0339,
|
|
"step": 2900
|
|
},
|
|
{
|
|
"epoch": 1.6799544419134396,
|
|
"grad_norm": 0.04736129939556122,
|
|
"learning_rate": 1.5211519687601692e-05,
|
|
"loss": 0.0358,
|
|
"step": 2950
|
|
},
|
|
{
|
|
"epoch": 1.7084282460136673,
|
|
"grad_norm": 0.17463508248329163,
|
|
"learning_rate": 1.5130165961601042e-05,
|
|
"loss": 0.034,
|
|
"step": 3000
|
|
},
|
|
{
|
|
"epoch": 1.7369020501138952,
|
|
"grad_norm": 0.058148205280303955,
|
|
"learning_rate": 1.5048812235600392e-05,
|
|
"loss": 0.0317,
|
|
"step": 3050
|
|
},
|
|
{
|
|
"epoch": 1.7653758542141231,
|
|
"grad_norm": 3.5262815952301025,
|
|
"learning_rate": 1.4967458509599741e-05,
|
|
"loss": 0.0477,
|
|
"step": 3100
|
|
},
|
|
{
|
|
"epoch": 1.7938496583143508,
|
|
"grad_norm": 2.971937894821167,
|
|
"learning_rate": 1.488610478359909e-05,
|
|
"loss": 0.0364,
|
|
"step": 3150
|
|
},
|
|
{
|
|
"epoch": 1.8223234624145785,
|
|
"grad_norm": 0.05181724950671196,
|
|
"learning_rate": 1.4804751057598439e-05,
|
|
"loss": 0.0394,
|
|
"step": 3200
|
|
},
|
|
{
|
|
"epoch": 1.8507972665148062,
|
|
"grad_norm": 0.2712370455265045,
|
|
"learning_rate": 1.4723397331597789e-05,
|
|
"loss": 0.0456,
|
|
"step": 3250
|
|
},
|
|
{
|
|
"epoch": 1.8792710706150342,
|
|
"grad_norm": 1.608622431755066,
|
|
"learning_rate": 1.4642043605597138e-05,
|
|
"loss": 0.05,
|
|
"step": 3300
|
|
},
|
|
{
|
|
"epoch": 1.907744874715262,
|
|
"grad_norm": 5.2423529624938965,
|
|
"learning_rate": 1.4560689879596486e-05,
|
|
"loss": 0.0371,
|
|
"step": 3350
|
|
},
|
|
{
|
|
"epoch": 1.9362186788154898,
|
|
"grad_norm": 0.04786211624741554,
|
|
"learning_rate": 1.4479336153595836e-05,
|
|
"loss": 0.0329,
|
|
"step": 3400
|
|
},
|
|
{
|
|
"epoch": 1.9646924829157175,
|
|
"grad_norm": 0.5264931917190552,
|
|
"learning_rate": 1.4397982427595185e-05,
|
|
"loss": 0.0379,
|
|
"step": 3450
|
|
},
|
|
{
|
|
"epoch": 1.9931662870159452,
|
|
"grad_norm": 0.236006498336792,
|
|
"learning_rate": 1.4316628701594535e-05,
|
|
"loss": 0.0253,
|
|
"step": 3500
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"eval_f1": 0.9406438631790744,
|
|
"eval_loss": 0.05223705992102623,
|
|
"eval_precision": 0.9371867691279653,
|
|
"eval_recall": 0.9441265567149109,
|
|
"eval_runtime": 3.7109,
|
|
"eval_samples_per_second": 875.791,
|
|
"eval_steps_per_second": 109.676,
|
|
"step": 3512
|
|
},
|
|
{
|
|
"epoch": 2.021640091116173,
|
|
"grad_norm": 4.225963115692139,
|
|
"learning_rate": 1.4235274975593883e-05,
|
|
"loss": 0.0373,
|
|
"step": 3550
|
|
},
|
|
{
|
|
"epoch": 2.050113895216401,
|
|
"grad_norm": 0.15388111770153046,
|
|
"learning_rate": 1.4153921249593233e-05,
|
|
"loss": 0.0258,
|
|
"step": 3600
|
|
},
|
|
{
|
|
"epoch": 2.078587699316629,
|
|
"grad_norm": 6.3911895751953125,
|
|
"learning_rate": 1.4072567523592582e-05,
|
|
"loss": 0.0217,
|
|
"step": 3650
|
|
},
|
|
{
|
|
"epoch": 2.1070615034168565,
|
|
"grad_norm": 0.03075530007481575,
|
|
"learning_rate": 1.3991213797591932e-05,
|
|
"loss": 0.019,
|
|
"step": 3700
|
|
},
|
|
{
|
|
"epoch": 2.135535307517084,
|
|
"grad_norm": 0.16612432897090912,
|
|
"learning_rate": 1.3909860071591278e-05,
|
|
"loss": 0.0175,
|
|
"step": 3750
|
|
},
|
|
{
|
|
"epoch": 2.164009111617312,
|
|
"grad_norm": 0.14543022215366364,
|
|
"learning_rate": 1.382850634559063e-05,
|
|
"loss": 0.0289,
|
|
"step": 3800
|
|
},
|
|
{
|
|
"epoch": 2.19248291571754,
|
|
"grad_norm": 2.9880359172821045,
|
|
"learning_rate": 1.374715261958998e-05,
|
|
"loss": 0.0208,
|
|
"step": 3850
|
|
},
|
|
{
|
|
"epoch": 2.2209567198177678,
|
|
"grad_norm": 9.878133773803711,
|
|
"learning_rate": 1.3665798893589329e-05,
|
|
"loss": 0.021,
|
|
"step": 3900
|
|
},
|
|
{
|
|
"epoch": 2.2494305239179955,
|
|
"grad_norm": 9.684176445007324,
|
|
"learning_rate": 1.3584445167588675e-05,
|
|
"loss": 0.0267,
|
|
"step": 3950
|
|
},
|
|
{
|
|
"epoch": 2.277904328018223,
|
|
"grad_norm": 33.15859603881836,
|
|
"learning_rate": 1.3503091441588025e-05,
|
|
"loss": 0.0295,
|
|
"step": 4000
|
|
},
|
|
{
|
|
"epoch": 2.306378132118451,
|
|
"grad_norm": 2.559544086456299,
|
|
"learning_rate": 1.3421737715587374e-05,
|
|
"loss": 0.0292,
|
|
"step": 4050
|
|
},
|
|
{
|
|
"epoch": 2.334851936218679,
|
|
"grad_norm": 0.07091552764177322,
|
|
"learning_rate": 1.3340383989586724e-05,
|
|
"loss": 0.0262,
|
|
"step": 4100
|
|
},
|
|
{
|
|
"epoch": 2.3633257403189067,
|
|
"grad_norm": 1.8119585514068604,
|
|
"learning_rate": 1.3259030263586072e-05,
|
|
"loss": 0.03,
|
|
"step": 4150
|
|
},
|
|
{
|
|
"epoch": 2.3917995444191344,
|
|
"grad_norm": 0.1476636379957199,
|
|
"learning_rate": 1.3177676537585422e-05,
|
|
"loss": 0.0196,
|
|
"step": 4200
|
|
},
|
|
{
|
|
"epoch": 2.420273348519362,
|
|
"grad_norm": 0.04482650011777878,
|
|
"learning_rate": 1.3096322811584771e-05,
|
|
"loss": 0.0263,
|
|
"step": 4250
|
|
},
|
|
{
|
|
"epoch": 2.44874715261959,
|
|
"grad_norm": 0.09201560169458389,
|
|
"learning_rate": 1.3014969085584121e-05,
|
|
"loss": 0.0172,
|
|
"step": 4300
|
|
},
|
|
{
|
|
"epoch": 2.477220956719818,
|
|
"grad_norm": 0.02215876244008541,
|
|
"learning_rate": 1.2933615359583469e-05,
|
|
"loss": 0.0149,
|
|
"step": 4350
|
|
},
|
|
{
|
|
"epoch": 2.5056947608200457,
|
|
"grad_norm": 0.6993932127952576,
|
|
"learning_rate": 1.2852261633582819e-05,
|
|
"loss": 0.0331,
|
|
"step": 4400
|
|
},
|
|
{
|
|
"epoch": 2.5341685649202734,
|
|
"grad_norm": 0.030453965067863464,
|
|
"learning_rate": 1.2770907907582168e-05,
|
|
"loss": 0.0307,
|
|
"step": 4450
|
|
},
|
|
{
|
|
"epoch": 2.562642369020501,
|
|
"grad_norm": 0.22211593389511108,
|
|
"learning_rate": 1.2689554181581518e-05,
|
|
"loss": 0.0183,
|
|
"step": 4500
|
|
},
|
|
{
|
|
"epoch": 2.591116173120729,
|
|
"grad_norm": 1.2842738628387451,
|
|
"learning_rate": 1.2608200455580866e-05,
|
|
"loss": 0.0226,
|
|
"step": 4550
|
|
},
|
|
{
|
|
"epoch": 2.619589977220957,
|
|
"grad_norm": 3.9048423767089844,
|
|
"learning_rate": 1.2526846729580216e-05,
|
|
"loss": 0.0239,
|
|
"step": 4600
|
|
},
|
|
{
|
|
"epoch": 2.6480637813211843,
|
|
"grad_norm": 0.9887784719467163,
|
|
"learning_rate": 1.2445493003579565e-05,
|
|
"loss": 0.0271,
|
|
"step": 4650
|
|
},
|
|
{
|
|
"epoch": 2.6765375854214124,
|
|
"grad_norm": 4.759728908538818,
|
|
"learning_rate": 1.2364139277578915e-05,
|
|
"loss": 0.0201,
|
|
"step": 4700
|
|
},
|
|
{
|
|
"epoch": 2.70501138952164,
|
|
"grad_norm": 0.2671191096305847,
|
|
"learning_rate": 1.2282785551578263e-05,
|
|
"loss": 0.0212,
|
|
"step": 4750
|
|
},
|
|
{
|
|
"epoch": 2.733485193621868,
|
|
"grad_norm": 0.036626849323511124,
|
|
"learning_rate": 1.2201431825577612e-05,
|
|
"loss": 0.0155,
|
|
"step": 4800
|
|
},
|
|
{
|
|
"epoch": 2.7619589977220955,
|
|
"grad_norm": 6.877408027648926,
|
|
"learning_rate": 1.2120078099576962e-05,
|
|
"loss": 0.0266,
|
|
"step": 4850
|
|
},
|
|
{
|
|
"epoch": 2.7904328018223232,
|
|
"grad_norm": 4.725896835327148,
|
|
"learning_rate": 1.2038724373576312e-05,
|
|
"loss": 0.0163,
|
|
"step": 4900
|
|
},
|
|
{
|
|
"epoch": 2.8189066059225514,
|
|
"grad_norm": 0.11965059489011765,
|
|
"learning_rate": 1.195737064757566e-05,
|
|
"loss": 0.0246,
|
|
"step": 4950
|
|
},
|
|
{
|
|
"epoch": 2.847380410022779,
|
|
"grad_norm": 0.6634079813957214,
|
|
"learning_rate": 1.187601692157501e-05,
|
|
"loss": 0.0304,
|
|
"step": 5000
|
|
},
|
|
{
|
|
"epoch": 2.875854214123007,
|
|
"grad_norm": 3.952694892883301,
|
|
"learning_rate": 1.1794663195574359e-05,
|
|
"loss": 0.0183,
|
|
"step": 5050
|
|
},
|
|
{
|
|
"epoch": 2.9043280182232345,
|
|
"grad_norm": 19.788942337036133,
|
|
"learning_rate": 1.1713309469573709e-05,
|
|
"loss": 0.0193,
|
|
"step": 5100
|
|
},
|
|
{
|
|
"epoch": 2.932801822323462,
|
|
"grad_norm": 4.989261627197266,
|
|
"learning_rate": 1.1631955743573055e-05,
|
|
"loss": 0.023,
|
|
"step": 5150
|
|
},
|
|
{
|
|
"epoch": 2.9612756264236904,
|
|
"grad_norm": 9.020853996276855,
|
|
"learning_rate": 1.1550602017572406e-05,
|
|
"loss": 0.0231,
|
|
"step": 5200
|
|
},
|
|
{
|
|
"epoch": 2.989749430523918,
|
|
"grad_norm": 0.9151140451431274,
|
|
"learning_rate": 1.1469248291571756e-05,
|
|
"loss": 0.0314,
|
|
"step": 5250
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"eval_f1": 0.9470420646891236,
|
|
"eval_loss": 0.04870199039578438,
|
|
"eval_precision": 0.943090787716956,
|
|
"eval_recall": 0.9510265903736116,
|
|
"eval_runtime": 3.7494,
|
|
"eval_samples_per_second": 866.805,
|
|
"eval_steps_per_second": 108.551,
|
|
"step": 5268
|
|
},
|
|
{
|
|
"epoch": 3.0182232346241458,
|
|
"grad_norm": 8.62992000579834,
|
|
"learning_rate": 1.1387894565571106e-05,
|
|
"loss": 0.0194,
|
|
"step": 5300
|
|
},
|
|
{
|
|
"epoch": 3.0466970387243735,
|
|
"grad_norm": 0.5615554451942444,
|
|
"learning_rate": 1.1306540839570452e-05,
|
|
"loss": 0.0114,
|
|
"step": 5350
|
|
},
|
|
{
|
|
"epoch": 3.075170842824601,
|
|
"grad_norm": 2.5735936164855957,
|
|
"learning_rate": 1.1225187113569801e-05,
|
|
"loss": 0.0163,
|
|
"step": 5400
|
|
},
|
|
{
|
|
"epoch": 3.1036446469248293,
|
|
"grad_norm": 6.504094123840332,
|
|
"learning_rate": 1.1143833387569151e-05,
|
|
"loss": 0.0135,
|
|
"step": 5450
|
|
},
|
|
{
|
|
"epoch": 3.132118451025057,
|
|
"grad_norm": 14.477537155151367,
|
|
"learning_rate": 1.10624796615685e-05,
|
|
"loss": 0.0129,
|
|
"step": 5500
|
|
},
|
|
{
|
|
"epoch": 3.1605922551252847,
|
|
"grad_norm": 0.04929841682314873,
|
|
"learning_rate": 1.0981125935567849e-05,
|
|
"loss": 0.0263,
|
|
"step": 5550
|
|
},
|
|
{
|
|
"epoch": 3.1890660592255125,
|
|
"grad_norm": 0.08875144273042679,
|
|
"learning_rate": 1.0899772209567198e-05,
|
|
"loss": 0.0134,
|
|
"step": 5600
|
|
},
|
|
{
|
|
"epoch": 3.21753986332574,
|
|
"grad_norm": 0.6238455772399902,
|
|
"learning_rate": 1.0818418483566548e-05,
|
|
"loss": 0.0107,
|
|
"step": 5650
|
|
},
|
|
{
|
|
"epoch": 3.2460136674259683,
|
|
"grad_norm": 0.004948179703205824,
|
|
"learning_rate": 1.0737064757565898e-05,
|
|
"loss": 0.0167,
|
|
"step": 5700
|
|
},
|
|
{
|
|
"epoch": 3.274487471526196,
|
|
"grad_norm": 0.017031218856573105,
|
|
"learning_rate": 1.0655711031565246e-05,
|
|
"loss": 0.0085,
|
|
"step": 5750
|
|
},
|
|
{
|
|
"epoch": 3.3029612756264237,
|
|
"grad_norm": 2.9840469360351562,
|
|
"learning_rate": 1.0574357305564595e-05,
|
|
"loss": 0.0158,
|
|
"step": 5800
|
|
},
|
|
{
|
|
"epoch": 3.3314350797266514,
|
|
"grad_norm": 1.7477540969848633,
|
|
"learning_rate": 1.0493003579563945e-05,
|
|
"loss": 0.0157,
|
|
"step": 5850
|
|
},
|
|
{
|
|
"epoch": 3.359908883826879,
|
|
"grad_norm": 0.03963543102145195,
|
|
"learning_rate": 1.0411649853563295e-05,
|
|
"loss": 0.0201,
|
|
"step": 5900
|
|
},
|
|
{
|
|
"epoch": 3.3883826879271073,
|
|
"grad_norm": 0.16669808328151703,
|
|
"learning_rate": 1.0330296127562643e-05,
|
|
"loss": 0.0159,
|
|
"step": 5950
|
|
},
|
|
{
|
|
"epoch": 3.416856492027335,
|
|
"grad_norm": 0.01953568309545517,
|
|
"learning_rate": 1.0248942401561992e-05,
|
|
"loss": 0.0117,
|
|
"step": 6000
|
|
},
|
|
{
|
|
"epoch": 3.4453302961275627,
|
|
"grad_norm": 0.07708246260881424,
|
|
"learning_rate": 1.0167588675561342e-05,
|
|
"loss": 0.0117,
|
|
"step": 6050
|
|
},
|
|
{
|
|
"epoch": 3.4738041002277904,
|
|
"grad_norm": 2.423590898513794,
|
|
"learning_rate": 1.0086234949560691e-05,
|
|
"loss": 0.0111,
|
|
"step": 6100
|
|
},
|
|
{
|
|
"epoch": 3.502277904328018,
|
|
"grad_norm": 0.006337775848805904,
|
|
"learning_rate": 1.000488122356004e-05,
|
|
"loss": 0.0052,
|
|
"step": 6150
|
|
},
|
|
{
|
|
"epoch": 3.5307517084282463,
|
|
"grad_norm": 0.01373753696680069,
|
|
"learning_rate": 9.923527497559389e-06,
|
|
"loss": 0.009,
|
|
"step": 6200
|
|
},
|
|
{
|
|
"epoch": 3.559225512528474,
|
|
"grad_norm": 0.11260247975587845,
|
|
"learning_rate": 9.842173771558739e-06,
|
|
"loss": 0.0208,
|
|
"step": 6250
|
|
},
|
|
{
|
|
"epoch": 3.5876993166287017,
|
|
"grad_norm": 1.240822196006775,
|
|
"learning_rate": 9.760820045558087e-06,
|
|
"loss": 0.0157,
|
|
"step": 6300
|
|
},
|
|
{
|
|
"epoch": 3.6161731207289294,
|
|
"grad_norm": 0.05271293595433235,
|
|
"learning_rate": 9.679466319557436e-06,
|
|
"loss": 0.0175,
|
|
"step": 6350
|
|
},
|
|
{
|
|
"epoch": 3.644646924829157,
|
|
"grad_norm": 24.546663284301758,
|
|
"learning_rate": 9.598112593556786e-06,
|
|
"loss": 0.015,
|
|
"step": 6400
|
|
},
|
|
{
|
|
"epoch": 3.6731207289293852,
|
|
"grad_norm": 5.120173931121826,
|
|
"learning_rate": 9.516758867556136e-06,
|
|
"loss": 0.017,
|
|
"step": 6450
|
|
},
|
|
{
|
|
"epoch": 3.7015945330296125,
|
|
"grad_norm": 1.1922008991241455,
|
|
"learning_rate": 9.435405141555484e-06,
|
|
"loss": 0.0194,
|
|
"step": 6500
|
|
},
|
|
{
|
|
"epoch": 3.7300683371298406,
|
|
"grad_norm": 0.039439987391233444,
|
|
"learning_rate": 9.354051415554833e-06,
|
|
"loss": 0.0309,
|
|
"step": 6550
|
|
},
|
|
{
|
|
"epoch": 3.7585421412300684,
|
|
"grad_norm": 0.04055279493331909,
|
|
"learning_rate": 9.272697689554181e-06,
|
|
"loss": 0.0195,
|
|
"step": 6600
|
|
},
|
|
{
|
|
"epoch": 3.787015945330296,
|
|
"grad_norm": 0.5971085429191589,
|
|
"learning_rate": 9.191343963553533e-06,
|
|
"loss": 0.0156,
|
|
"step": 6650
|
|
},
|
|
{
|
|
"epoch": 3.8154897494305238,
|
|
"grad_norm": 0.2191866636276245,
|
|
"learning_rate": 9.10999023755288e-06,
|
|
"loss": 0.0174,
|
|
"step": 6700
|
|
},
|
|
{
|
|
"epoch": 3.8439635535307515,
|
|
"grad_norm": 0.033559828996658325,
|
|
"learning_rate": 9.02863651155223e-06,
|
|
"loss": 0.0149,
|
|
"step": 6750
|
|
},
|
|
{
|
|
"epoch": 3.8724373576309796,
|
|
"grad_norm": 3.7128634452819824,
|
|
"learning_rate": 8.947282785551578e-06,
|
|
"loss": 0.0221,
|
|
"step": 6800
|
|
},
|
|
{
|
|
"epoch": 3.9009111617312073,
|
|
"grad_norm": 0.08547580987215042,
|
|
"learning_rate": 8.865929059550928e-06,
|
|
"loss": 0.0153,
|
|
"step": 6850
|
|
},
|
|
{
|
|
"epoch": 3.929384965831435,
|
|
"grad_norm": 0.058124080300331116,
|
|
"learning_rate": 8.784575333550277e-06,
|
|
"loss": 0.0161,
|
|
"step": 6900
|
|
},
|
|
{
|
|
"epoch": 3.9578587699316627,
|
|
"grad_norm": 0.015579139813780785,
|
|
"learning_rate": 8.703221607549627e-06,
|
|
"loss": 0.011,
|
|
"step": 6950
|
|
},
|
|
{
|
|
"epoch": 3.9863325740318905,
|
|
"grad_norm": 0.12764935195446014,
|
|
"learning_rate": 8.621867881548975e-06,
|
|
"loss": 0.0166,
|
|
"step": 7000
|
|
},
|
|
{
|
|
"epoch": 4.0,
|
|
"eval_f1": 0.9525329748802823,
|
|
"eval_loss": 0.04983380436897278,
|
|
"eval_precision": 0.9510149303808086,
|
|
"eval_recall": 0.9540558734432851,
|
|
"eval_runtime": 3.7495,
|
|
"eval_samples_per_second": 866.79,
|
|
"eval_steps_per_second": 108.549,
|
|
"step": 7024
|
|
},
|
|
{
|
|
"epoch": 4.014806378132119,
|
|
"grad_norm": 0.04521024227142334,
|
|
"learning_rate": 8.540514155548325e-06,
|
|
"loss": 0.0102,
|
|
"step": 7050
|
|
},
|
|
{
|
|
"epoch": 4.043280182232346,
|
|
"grad_norm": 0.015535669401288033,
|
|
"learning_rate": 8.459160429547674e-06,
|
|
"loss": 0.0068,
|
|
"step": 7100
|
|
},
|
|
{
|
|
"epoch": 4.071753986332574,
|
|
"grad_norm": 12.566010475158691,
|
|
"learning_rate": 8.377806703547024e-06,
|
|
"loss": 0.0151,
|
|
"step": 7150
|
|
},
|
|
{
|
|
"epoch": 4.100227790432802,
|
|
"grad_norm": 0.027761396020650864,
|
|
"learning_rate": 8.296452977546372e-06,
|
|
"loss": 0.0053,
|
|
"step": 7200
|
|
},
|
|
{
|
|
"epoch": 4.128701594533029,
|
|
"grad_norm": 1.8846938610076904,
|
|
"learning_rate": 8.215099251545722e-06,
|
|
"loss": 0.0077,
|
|
"step": 7250
|
|
},
|
|
{
|
|
"epoch": 4.157175398633258,
|
|
"grad_norm": 0.005013479385524988,
|
|
"learning_rate": 8.13374552554507e-06,
|
|
"loss": 0.0073,
|
|
"step": 7300
|
|
},
|
|
{
|
|
"epoch": 4.185649202733485,
|
|
"grad_norm": 0.32696235179901123,
|
|
"learning_rate": 8.052391799544421e-06,
|
|
"loss": 0.0092,
|
|
"step": 7350
|
|
},
|
|
{
|
|
"epoch": 4.214123006833713,
|
|
"grad_norm": 0.014357910491526127,
|
|
"learning_rate": 7.971038073543769e-06,
|
|
"loss": 0.0164,
|
|
"step": 7400
|
|
},
|
|
{
|
|
"epoch": 4.242596810933941,
|
|
"grad_norm": 0.010845329612493515,
|
|
"learning_rate": 7.889684347543118e-06,
|
|
"loss": 0.0123,
|
|
"step": 7450
|
|
},
|
|
{
|
|
"epoch": 4.271070615034168,
|
|
"grad_norm": 0.6969354748725891,
|
|
"learning_rate": 7.808330621542466e-06,
|
|
"loss": 0.0103,
|
|
"step": 7500
|
|
},
|
|
{
|
|
"epoch": 4.2995444191343966,
|
|
"grad_norm": 0.3575742840766907,
|
|
"learning_rate": 7.726976895541816e-06,
|
|
"loss": 0.0161,
|
|
"step": 7550
|
|
},
|
|
{
|
|
"epoch": 4.328018223234624,
|
|
"grad_norm": 0.008076228201389313,
|
|
"learning_rate": 7.645623169541166e-06,
|
|
"loss": 0.0122,
|
|
"step": 7600
|
|
},
|
|
{
|
|
"epoch": 4.356492027334852,
|
|
"grad_norm": 0.029101597145199776,
|
|
"learning_rate": 7.564269443540515e-06,
|
|
"loss": 0.0071,
|
|
"step": 7650
|
|
},
|
|
{
|
|
"epoch": 4.38496583143508,
|
|
"grad_norm": 0.06841142475605011,
|
|
"learning_rate": 7.482915717539863e-06,
|
|
"loss": 0.0091,
|
|
"step": 7700
|
|
},
|
|
{
|
|
"epoch": 4.413439635535307,
|
|
"grad_norm": 0.04133535176515579,
|
|
"learning_rate": 7.401561991539213e-06,
|
|
"loss": 0.0073,
|
|
"step": 7750
|
|
},
|
|
{
|
|
"epoch": 4.4419134396355355,
|
|
"grad_norm": 0.03571132943034172,
|
|
"learning_rate": 7.320208265538562e-06,
|
|
"loss": 0.0131,
|
|
"step": 7800
|
|
},
|
|
{
|
|
"epoch": 4.470387243735763,
|
|
"grad_norm": 2.8286924362182617,
|
|
"learning_rate": 7.2388545395379114e-06,
|
|
"loss": 0.0043,
|
|
"step": 7850
|
|
},
|
|
{
|
|
"epoch": 4.498861047835991,
|
|
"grad_norm": 1.3785158395767212,
|
|
"learning_rate": 7.15750081353726e-06,
|
|
"loss": 0.0103,
|
|
"step": 7900
|
|
},
|
|
{
|
|
"epoch": 4.527334851936219,
|
|
"grad_norm": 0.05979786813259125,
|
|
"learning_rate": 7.07614708753661e-06,
|
|
"loss": 0.006,
|
|
"step": 7950
|
|
},
|
|
{
|
|
"epoch": 4.555808656036446,
|
|
"grad_norm": 1.3434393405914307,
|
|
"learning_rate": 6.994793361535959e-06,
|
|
"loss": 0.0148,
|
|
"step": 8000
|
|
},
|
|
{
|
|
"epoch": 4.5842824601366745,
|
|
"grad_norm": 0.019605603069067,
|
|
"learning_rate": 6.913439635535308e-06,
|
|
"loss": 0.0068,
|
|
"step": 8050
|
|
},
|
|
{
|
|
"epoch": 4.612756264236902,
|
|
"grad_norm": 0.0028891051188111305,
|
|
"learning_rate": 6.832085909534657e-06,
|
|
"loss": 0.0075,
|
|
"step": 8100
|
|
},
|
|
{
|
|
"epoch": 4.64123006833713,
|
|
"grad_norm": 0.014856363646686077,
|
|
"learning_rate": 6.750732183534007e-06,
|
|
"loss": 0.0073,
|
|
"step": 8150
|
|
},
|
|
{
|
|
"epoch": 4.669703872437358,
|
|
"grad_norm": 0.0037149768322706223,
|
|
"learning_rate": 6.669378457533356e-06,
|
|
"loss": 0.0066,
|
|
"step": 8200
|
|
},
|
|
{
|
|
"epoch": 4.698177676537585,
|
|
"grad_norm": 0.008534993045032024,
|
|
"learning_rate": 6.588024731532705e-06,
|
|
"loss": 0.0121,
|
|
"step": 8250
|
|
},
|
|
{
|
|
"epoch": 4.7266514806378135,
|
|
"grad_norm": 0.013920712284743786,
|
|
"learning_rate": 6.506671005532054e-06,
|
|
"loss": 0.013,
|
|
"step": 8300
|
|
},
|
|
{
|
|
"epoch": 4.755125284738041,
|
|
"grad_norm": 0.10121840238571167,
|
|
"learning_rate": 6.425317279531404e-06,
|
|
"loss": 0.0087,
|
|
"step": 8350
|
|
},
|
|
{
|
|
"epoch": 4.783599088838269,
|
|
"grad_norm": 0.07937771081924438,
|
|
"learning_rate": 6.343963553530752e-06,
|
|
"loss": 0.0086,
|
|
"step": 8400
|
|
},
|
|
{
|
|
"epoch": 4.812072892938497,
|
|
"grad_norm": 0.07674901187419891,
|
|
"learning_rate": 6.262609827530101e-06,
|
|
"loss": 0.0052,
|
|
"step": 8450
|
|
},
|
|
{
|
|
"epoch": 4.840546697038724,
|
|
"grad_norm": 0.06683178246021271,
|
|
"learning_rate": 6.18125610152945e-06,
|
|
"loss": 0.015,
|
|
"step": 8500
|
|
},
|
|
{
|
|
"epoch": 4.8690205011389525,
|
|
"grad_norm": 0.08861212432384491,
|
|
"learning_rate": 6.0999023755288e-06,
|
|
"loss": 0.0109,
|
|
"step": 8550
|
|
},
|
|
{
|
|
"epoch": 4.89749430523918,
|
|
"grad_norm": 0.006198623217642307,
|
|
"learning_rate": 6.018548649528149e-06,
|
|
"loss": 0.0181,
|
|
"step": 8600
|
|
},
|
|
{
|
|
"epoch": 4.925968109339408,
|
|
"grad_norm": 0.009029284119606018,
|
|
"learning_rate": 5.937194923527498e-06,
|
|
"loss": 0.0043,
|
|
"step": 8650
|
|
},
|
|
{
|
|
"epoch": 4.954441913439636,
|
|
"grad_norm": 1.1225602626800537,
|
|
"learning_rate": 5.855841197526847e-06,
|
|
"loss": 0.0121,
|
|
"step": 8700
|
|
},
|
|
{
|
|
"epoch": 4.982915717539863,
|
|
"grad_norm": 0.01218325924128294,
|
|
"learning_rate": 5.774487471526197e-06,
|
|
"loss": 0.0115,
|
|
"step": 8750
|
|
},
|
|
{
|
|
"epoch": 5.0,
|
|
"eval_f1": 0.9533444816053511,
|
|
"eval_loss": 0.04722925275564194,
|
|
"eval_precision": 0.9473246925889,
|
|
"eval_recall": 0.9594412655671491,
|
|
"eval_runtime": 3.7358,
|
|
"eval_samples_per_second": 869.968,
|
|
"eval_steps_per_second": 108.947,
|
|
"step": 8780
|
|
},
|
|
{
|
|
"epoch": 5.011389521640091,
|
|
"grad_norm": 0.8161097764968872,
|
|
"learning_rate": 5.6931337455255455e-06,
|
|
"loss": 0.0127,
|
|
"step": 8800
|
|
},
|
|
{
|
|
"epoch": 5.039863325740319,
|
|
"grad_norm": 0.048089127987623215,
|
|
"learning_rate": 5.611780019524895e-06,
|
|
"loss": 0.0116,
|
|
"step": 8850
|
|
},
|
|
{
|
|
"epoch": 5.068337129840547,
|
|
"grad_norm": 1.7576072216033936,
|
|
"learning_rate": 5.530426293524244e-06,
|
|
"loss": 0.0093,
|
|
"step": 8900
|
|
},
|
|
{
|
|
"epoch": 5.096810933940774,
|
|
"grad_norm": 0.07270015776157379,
|
|
"learning_rate": 5.449072567523594e-06,
|
|
"loss": 0.0043,
|
|
"step": 8950
|
|
},
|
|
{
|
|
"epoch": 5.125284738041002,
|
|
"grad_norm": 0.2580782175064087,
|
|
"learning_rate": 5.367718841522942e-06,
|
|
"loss": 0.0058,
|
|
"step": 9000
|
|
},
|
|
{
|
|
"epoch": 5.15375854214123,
|
|
"grad_norm": 0.12649740278720856,
|
|
"learning_rate": 5.286365115522292e-06,
|
|
"loss": 0.0033,
|
|
"step": 9050
|
|
},
|
|
{
|
|
"epoch": 5.182232346241458,
|
|
"grad_norm": 0.06850716471672058,
|
|
"learning_rate": 5.20501138952164e-06,
|
|
"loss": 0.003,
|
|
"step": 9100
|
|
},
|
|
{
|
|
"epoch": 5.210706150341686,
|
|
"grad_norm": 0.01197959017008543,
|
|
"learning_rate": 5.12365766352099e-06,
|
|
"loss": 0.0029,
|
|
"step": 9150
|
|
},
|
|
{
|
|
"epoch": 5.239179954441913,
|
|
"grad_norm": 0.10818086564540863,
|
|
"learning_rate": 5.0423039375203385e-06,
|
|
"loss": 0.0025,
|
|
"step": 9200
|
|
},
|
|
{
|
|
"epoch": 5.267653758542141,
|
|
"grad_norm": 0.035460665822029114,
|
|
"learning_rate": 4.960950211519688e-06,
|
|
"loss": 0.0047,
|
|
"step": 9250
|
|
},
|
|
{
|
|
"epoch": 5.296127562642369,
|
|
"grad_norm": 4.532562255859375,
|
|
"learning_rate": 4.879596485519037e-06,
|
|
"loss": 0.0086,
|
|
"step": 9300
|
|
},
|
|
{
|
|
"epoch": 5.324601366742597,
|
|
"grad_norm": 0.013236219063401222,
|
|
"learning_rate": 4.7982427595183866e-06,
|
|
"loss": 0.0027,
|
|
"step": 9350
|
|
},
|
|
{
|
|
"epoch": 5.353075170842825,
|
|
"grad_norm": 0.018059909343719482,
|
|
"learning_rate": 4.716889033517735e-06,
|
|
"loss": 0.0065,
|
|
"step": 9400
|
|
},
|
|
{
|
|
"epoch": 5.381548974943052,
|
|
"grad_norm": 0.019096272066235542,
|
|
"learning_rate": 4.635535307517084e-06,
|
|
"loss": 0.0053,
|
|
"step": 9450
|
|
},
|
|
{
|
|
"epoch": 5.41002277904328,
|
|
"grad_norm": 0.10389436781406403,
|
|
"learning_rate": 4.554181581516434e-06,
|
|
"loss": 0.0029,
|
|
"step": 9500
|
|
},
|
|
{
|
|
"epoch": 5.438496583143508,
|
|
"grad_norm": 0.0029252381063997746,
|
|
"learning_rate": 4.472827855515783e-06,
|
|
"loss": 0.0045,
|
|
"step": 9550
|
|
},
|
|
{
|
|
"epoch": 5.466970387243736,
|
|
"grad_norm": 0.004542892333120108,
|
|
"learning_rate": 4.391474129515132e-06,
|
|
"loss": 0.0102,
|
|
"step": 9600
|
|
},
|
|
{
|
|
"epoch": 5.495444191343964,
|
|
"grad_norm": 0.130916565656662,
|
|
"learning_rate": 4.310120403514481e-06,
|
|
"loss": 0.0064,
|
|
"step": 9650
|
|
},
|
|
{
|
|
"epoch": 5.523917995444191,
|
|
"grad_norm": 0.008013393729925156,
|
|
"learning_rate": 4.22876667751383e-06,
|
|
"loss": 0.005,
|
|
"step": 9700
|
|
},
|
|
{
|
|
"epoch": 5.552391799544419,
|
|
"grad_norm": 0.012782514095306396,
|
|
"learning_rate": 4.1474129515131795e-06,
|
|
"loss": 0.0122,
|
|
"step": 9750
|
|
},
|
|
{
|
|
"epoch": 5.5808656036446465,
|
|
"grad_norm": 0.9617053866386414,
|
|
"learning_rate": 4.066059225512528e-06,
|
|
"loss": 0.0063,
|
|
"step": 9800
|
|
},
|
|
{
|
|
"epoch": 5.609339407744875,
|
|
"grad_norm": 0.014953136444091797,
|
|
"learning_rate": 3.984705499511878e-06,
|
|
"loss": 0.0046,
|
|
"step": 9850
|
|
},
|
|
{
|
|
"epoch": 5.637813211845103,
|
|
"grad_norm": 1.0275633335113525,
|
|
"learning_rate": 3.903351773511227e-06,
|
|
"loss": 0.0075,
|
|
"step": 9900
|
|
},
|
|
{
|
|
"epoch": 5.66628701594533,
|
|
"grad_norm": 0.007999264635145664,
|
|
"learning_rate": 3.8219980475105764e-06,
|
|
"loss": 0.0031,
|
|
"step": 9950
|
|
},
|
|
{
|
|
"epoch": 5.694760820045558,
|
|
"grad_norm": 0.008241960778832436,
|
|
"learning_rate": 3.7406443215099252e-06,
|
|
"loss": 0.0065,
|
|
"step": 10000
|
|
},
|
|
{
|
|
"epoch": 5.723234624145785,
|
|
"grad_norm": 2.578386068344116,
|
|
"learning_rate": 3.6592905955092745e-06,
|
|
"loss": 0.0057,
|
|
"step": 10050
|
|
},
|
|
{
|
|
"epoch": 5.751708428246014,
|
|
"grad_norm": 2.8855443000793457,
|
|
"learning_rate": 3.5779368695086237e-06,
|
|
"loss": 0.0038,
|
|
"step": 10100
|
|
},
|
|
{
|
|
"epoch": 5.780182232346242,
|
|
"grad_norm": 0.016262667253613472,
|
|
"learning_rate": 3.496583143507973e-06,
|
|
"loss": 0.0128,
|
|
"step": 10150
|
|
},
|
|
{
|
|
"epoch": 5.808656036446469,
|
|
"grad_norm": 0.045168060809373856,
|
|
"learning_rate": 3.415229417507322e-06,
|
|
"loss": 0.0036,
|
|
"step": 10200
|
|
},
|
|
{
|
|
"epoch": 5.837129840546697,
|
|
"grad_norm": 0.1192622259259224,
|
|
"learning_rate": 3.3355027660266842e-06,
|
|
"loss": 0.0111,
|
|
"step": 10250
|
|
},
|
|
{
|
|
"epoch": 5.865603644646924,
|
|
"grad_norm": 0.08544190227985382,
|
|
"learning_rate": 3.2541490400260335e-06,
|
|
"loss": 0.0051,
|
|
"step": 10300
|
|
},
|
|
{
|
|
"epoch": 5.894077448747153,
|
|
"grad_norm": 3.493161678314209,
|
|
"learning_rate": 3.1727953140253827e-06,
|
|
"loss": 0.0054,
|
|
"step": 10350
|
|
},
|
|
{
|
|
"epoch": 5.922551252847381,
|
|
"grad_norm": 0.022789066657423973,
|
|
"learning_rate": 3.0914415880247315e-06,
|
|
"loss": 0.0033,
|
|
"step": 10400
|
|
},
|
|
{
|
|
"epoch": 5.951025056947608,
|
|
"grad_norm": 12.131625175476074,
|
|
"learning_rate": 3.0100878620240807e-06,
|
|
"loss": 0.0023,
|
|
"step": 10450
|
|
},
|
|
{
|
|
"epoch": 5.979498861047836,
|
|
"grad_norm": 0.03329641371965408,
|
|
"learning_rate": 2.92873413602343e-06,
|
|
"loss": 0.0058,
|
|
"step": 10500
|
|
},
|
|
{
|
|
"epoch": 6.0,
|
|
"eval_f1": 0.9557566616390145,
|
|
"eval_loss": 0.05410688370466232,
|
|
"eval_precision": 0.9517690253671562,
|
|
"eval_recall": 0.9597778525748906,
|
|
"eval_runtime": 3.6572,
|
|
"eval_samples_per_second": 888.662,
|
|
"eval_steps_per_second": 111.288,
|
|
"step": 10536
|
|
},
|
|
{
|
|
"epoch": 6.007972665148063,
|
|
"grad_norm": 0.008557640947401524,
|
|
"learning_rate": 2.847380410022779e-06,
|
|
"loss": 0.0152,
|
|
"step": 10550
|
|
},
|
|
{
|
|
"epoch": 6.0364464692482915,
|
|
"grad_norm": 0.08105529844760895,
|
|
"learning_rate": 2.7660266840221284e-06,
|
|
"loss": 0.0025,
|
|
"step": 10600
|
|
},
|
|
{
|
|
"epoch": 6.06492027334852,
|
|
"grad_norm": 1.100066065788269,
|
|
"learning_rate": 2.6846729580214776e-06,
|
|
"loss": 0.0015,
|
|
"step": 10650
|
|
},
|
|
{
|
|
"epoch": 6.093394077448747,
|
|
"grad_norm": 1.96909761428833,
|
|
"learning_rate": 2.603319232020827e-06,
|
|
"loss": 0.0021,
|
|
"step": 10700
|
|
},
|
|
{
|
|
"epoch": 6.121867881548975,
|
|
"grad_norm": 0.00583766121417284,
|
|
"learning_rate": 2.5219655060201757e-06,
|
|
"loss": 0.0127,
|
|
"step": 10750
|
|
},
|
|
{
|
|
"epoch": 6.150341685649202,
|
|
"grad_norm": 0.026862677186727524,
|
|
"learning_rate": 2.440611780019525e-06,
|
|
"loss": 0.004,
|
|
"step": 10800
|
|
},
|
|
{
|
|
"epoch": 6.1788154897494305,
|
|
"grad_norm": 0.010042566806077957,
|
|
"learning_rate": 2.359258054018874e-06,
|
|
"loss": 0.0033,
|
|
"step": 10850
|
|
},
|
|
{
|
|
"epoch": 6.207289293849659,
|
|
"grad_norm": 0.8956929445266724,
|
|
"learning_rate": 2.2779043280182233e-06,
|
|
"loss": 0.0031,
|
|
"step": 10900
|
|
},
|
|
{
|
|
"epoch": 6.235763097949886,
|
|
"grad_norm": 0.009118441492319107,
|
|
"learning_rate": 2.1965506020175726e-06,
|
|
"loss": 0.0039,
|
|
"step": 10950
|
|
},
|
|
{
|
|
"epoch": 6.264236902050114,
|
|
"grad_norm": 0.22793921828269958,
|
|
"learning_rate": 2.115196876016922e-06,
|
|
"loss": 0.0013,
|
|
"step": 11000
|
|
},
|
|
{
|
|
"epoch": 6.292710706150341,
|
|
"grad_norm": 0.015608682297170162,
|
|
"learning_rate": 2.033843150016271e-06,
|
|
"loss": 0.0021,
|
|
"step": 11050
|
|
},
|
|
{
|
|
"epoch": 6.3211845102505695,
|
|
"grad_norm": 0.004031027667224407,
|
|
"learning_rate": 1.95248942401562e-06,
|
|
"loss": 0.0011,
|
|
"step": 11100
|
|
},
|
|
{
|
|
"epoch": 6.349658314350798,
|
|
"grad_norm": 0.008949857205152512,
|
|
"learning_rate": 1.8711356980149693e-06,
|
|
"loss": 0.0049,
|
|
"step": 11150
|
|
},
|
|
{
|
|
"epoch": 6.378132118451025,
|
|
"grad_norm": 0.018670039251446724,
|
|
"learning_rate": 1.7897819720143183e-06,
|
|
"loss": 0.003,
|
|
"step": 11200
|
|
},
|
|
{
|
|
"epoch": 6.406605922551253,
|
|
"grad_norm": 0.032393742352724075,
|
|
"learning_rate": 1.7084282460136675e-06,
|
|
"loss": 0.0051,
|
|
"step": 11250
|
|
},
|
|
{
|
|
"epoch": 6.43507972665148,
|
|
"grad_norm": 0.11851054430007935,
|
|
"learning_rate": 1.6270745200130167e-06,
|
|
"loss": 0.0046,
|
|
"step": 11300
|
|
},
|
|
{
|
|
"epoch": 6.4635535307517085,
|
|
"grad_norm": 0.09247086197137833,
|
|
"learning_rate": 1.5457207940123657e-06,
|
|
"loss": 0.0049,
|
|
"step": 11350
|
|
},
|
|
{
|
|
"epoch": 6.492027334851937,
|
|
"grad_norm": 0.002555207349359989,
|
|
"learning_rate": 1.464367068011715e-06,
|
|
"loss": 0.0026,
|
|
"step": 11400
|
|
},
|
|
{
|
|
"epoch": 6.520501138952164,
|
|
"grad_norm": 3.7294840812683105,
|
|
"learning_rate": 1.3830133420110642e-06,
|
|
"loss": 0.0014,
|
|
"step": 11450
|
|
},
|
|
{
|
|
"epoch": 6.548974943052392,
|
|
"grad_norm": 0.00703430688008666,
|
|
"learning_rate": 1.3016596160104134e-06,
|
|
"loss": 0.0035,
|
|
"step": 11500
|
|
},
|
|
{
|
|
"epoch": 6.577448747152619,
|
|
"grad_norm": 12.061240196228027,
|
|
"learning_rate": 1.2203058900097624e-06,
|
|
"loss": 0.0078,
|
|
"step": 11550
|
|
},
|
|
{
|
|
"epoch": 6.605922551252847,
|
|
"grad_norm": 0.011964640580117702,
|
|
"learning_rate": 1.1389521640091117e-06,
|
|
"loss": 0.0024,
|
|
"step": 11600
|
|
},
|
|
{
|
|
"epoch": 6.634396355353076,
|
|
"grad_norm": 0.014603933319449425,
|
|
"learning_rate": 1.057598438008461e-06,
|
|
"loss": 0.0072,
|
|
"step": 11650
|
|
},
|
|
{
|
|
"epoch": 6.662870159453303,
|
|
"grad_norm": 0.026642296463251114,
|
|
"learning_rate": 9.7624471200781e-07,
|
|
"loss": 0.0007,
|
|
"step": 11700
|
|
},
|
|
{
|
|
"epoch": 6.691343963553531,
|
|
"grad_norm": 3.3646230697631836,
|
|
"learning_rate": 8.948909860071591e-07,
|
|
"loss": 0.0046,
|
|
"step": 11750
|
|
},
|
|
{
|
|
"epoch": 6.719817767653758,
|
|
"grad_norm": 0.027053840458393097,
|
|
"learning_rate": 8.135372600065084e-07,
|
|
"loss": 0.0031,
|
|
"step": 11800
|
|
},
|
|
{
|
|
"epoch": 6.748291571753986,
|
|
"grad_norm": 2.6146697998046875,
|
|
"learning_rate": 7.321835340058575e-07,
|
|
"loss": 0.003,
|
|
"step": 11850
|
|
},
|
|
{
|
|
"epoch": 6.776765375854215,
|
|
"grad_norm": 8.964512825012207,
|
|
"learning_rate": 6.508298080052067e-07,
|
|
"loss": 0.0069,
|
|
"step": 11900
|
|
},
|
|
{
|
|
"epoch": 6.805239179954442,
|
|
"grad_norm": 0.3086203932762146,
|
|
"learning_rate": 5.694760820045558e-07,
|
|
"loss": 0.0035,
|
|
"step": 11950
|
|
},
|
|
{
|
|
"epoch": 6.83371298405467,
|
|
"grad_norm": 0.018831729888916016,
|
|
"learning_rate": 4.88122356003905e-07,
|
|
"loss": 0.0033,
|
|
"step": 12000
|
|
},
|
|
{
|
|
"epoch": 6.862186788154897,
|
|
"grad_norm": 0.0018483272287994623,
|
|
"learning_rate": 4.067686300032542e-07,
|
|
"loss": 0.0022,
|
|
"step": 12050
|
|
},
|
|
{
|
|
"epoch": 6.890660592255125,
|
|
"grad_norm": 0.003242627950385213,
|
|
"learning_rate": 3.2541490400260336e-07,
|
|
"loss": 0.0043,
|
|
"step": 12100
|
|
},
|
|
{
|
|
"epoch": 6.9191343963553535,
|
|
"grad_norm": 0.002707740291953087,
|
|
"learning_rate": 2.440611780019525e-07,
|
|
"loss": 0.0017,
|
|
"step": 12150
|
|
},
|
|
{
|
|
"epoch": 6.947608200455581,
|
|
"grad_norm": 0.21824024617671967,
|
|
"learning_rate": 1.6270745200130168e-07,
|
|
"loss": 0.0069,
|
|
"step": 12200
|
|
},
|
|
{
|
|
"epoch": 6.976082004555809,
|
|
"grad_norm": 0.008130647242069244,
|
|
"learning_rate": 8.135372600065084e-08,
|
|
"loss": 0.0047,
|
|
"step": 12250
|
|
},
|
|
{
|
|
"epoch": 7.0,
|
|
"eval_f1": 0.9563834240267894,
|
|
"eval_loss": 0.052909377962350845,
|
|
"eval_precision": 0.9515242378810594,
|
|
"eval_recall": 0.9612924941097274,
|
|
"eval_runtime": 3.7426,
|
|
"eval_samples_per_second": 868.392,
|
|
"eval_steps_per_second": 108.749,
|
|
"step": 12292
|
|
}
|
|
],
|
|
"logging_steps": 50,
|
|
"max_steps": 12292,
|
|
"num_input_tokens_seen": 0,
|
|
"num_train_epochs": 7,
|
|
"save_steps": 500,
|
|
"stateful_callbacks": {
|
|
"TrainerControl": {
|
|
"args": {
|
|
"should_epoch_stop": false,
|
|
"should_evaluate": false,
|
|
"should_log": false,
|
|
"should_save": true,
|
|
"should_training_stop": true
|
|
},
|
|
"attributes": {}
|
|
}
|
|
},
|
|
"total_flos": 2418024281348142.0,
|
|
"train_batch_size": 8,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|
|
|