oodeh's picture
Add files using upload-large-folder tool
a60a2a0 verified
raw
history blame
12.3 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9908256880733946,
"eval_steps": 500,
"global_step": 54,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01834862385321101,
"grad_norm": 0.04378490149974823,
"learning_rate": 4.999989423013716e-05,
"loss": 0.6713,
"num_input_tokens_seen": 44136,
"step": 1
},
{
"epoch": 0.03669724770642202,
"grad_norm": 0.040646992623806,
"learning_rate": 4.999957692144361e-05,
"loss": 0.533,
"num_input_tokens_seen": 83096,
"step": 2
},
{
"epoch": 0.05504587155963303,
"grad_norm": 0.04658753052353859,
"learning_rate": 4.999904807660428e-05,
"loss": 0.6048,
"num_input_tokens_seen": 122112,
"step": 3
},
{
"epoch": 0.07339449541284404,
"grad_norm": 0.04322144016623497,
"learning_rate": 4.999830770009406e-05,
"loss": 0.4948,
"num_input_tokens_seen": 163064,
"step": 4
},
{
"epoch": 0.09174311926605505,
"grad_norm": 0.06536195427179337,
"learning_rate": 4.999735579817769e-05,
"loss": 0.6607,
"num_input_tokens_seen": 203808,
"step": 5
},
{
"epoch": 0.11009174311926606,
"grad_norm": 0.059904925525188446,
"learning_rate": 4.9996192378909786e-05,
"loss": 0.5802,
"num_input_tokens_seen": 241824,
"step": 6
},
{
"epoch": 0.12844036697247707,
"grad_norm": 0.19818365573883057,
"learning_rate": 4.999481745213471e-05,
"loss": 0.5148,
"num_input_tokens_seen": 287608,
"step": 7
},
{
"epoch": 0.14678899082568808,
"grad_norm": 0.05985472351312637,
"learning_rate": 4.9993231029486544e-05,
"loss": 0.5714,
"num_input_tokens_seen": 325320,
"step": 8
},
{
"epoch": 0.1651376146788991,
"grad_norm": 0.061375778168439865,
"learning_rate": 4.999143312438893e-05,
"loss": 0.6812,
"num_input_tokens_seen": 369848,
"step": 9
},
{
"epoch": 0.1834862385321101,
"grad_norm": 0.06196414306759834,
"learning_rate": 4.998942375205502e-05,
"loss": 0.5358,
"num_input_tokens_seen": 415104,
"step": 10
},
{
"epoch": 0.2018348623853211,
"grad_norm": 0.07861393690109253,
"learning_rate": 4.9987202929487275e-05,
"loss": 0.6527,
"num_input_tokens_seen": 467224,
"step": 11
},
{
"epoch": 0.22018348623853212,
"grad_norm": 0.05596446990966797,
"learning_rate": 4.99847706754774e-05,
"loss": 0.5354,
"num_input_tokens_seen": 502824,
"step": 12
},
{
"epoch": 0.23853211009174313,
"grad_norm": 0.05289844051003456,
"learning_rate": 4.998212701060612e-05,
"loss": 0.5263,
"num_input_tokens_seen": 544744,
"step": 13
},
{
"epoch": 0.25688073394495414,
"grad_norm": 0.04996591433882713,
"learning_rate": 4.997927195724303e-05,
"loss": 0.5536,
"num_input_tokens_seen": 591136,
"step": 14
},
{
"epoch": 0.27522935779816515,
"grad_norm": 0.05822828412055969,
"learning_rate": 4.997620553954645e-05,
"loss": 0.6106,
"num_input_tokens_seen": 629664,
"step": 15
},
{
"epoch": 0.29357798165137616,
"grad_norm": 0.06353770196437836,
"learning_rate": 4.997292778346312e-05,
"loss": 0.5129,
"num_input_tokens_seen": 663392,
"step": 16
},
{
"epoch": 0.3119266055045872,
"grad_norm": 0.07256966829299927,
"learning_rate": 4.996943871672807e-05,
"loss": 0.6377,
"num_input_tokens_seen": 698360,
"step": 17
},
{
"epoch": 0.3302752293577982,
"grad_norm": 0.055458713322877884,
"learning_rate": 4.996573836886435e-05,
"loss": 0.4083,
"num_input_tokens_seen": 737520,
"step": 18
},
{
"epoch": 0.3486238532110092,
"grad_norm": 0.07792335003614426,
"learning_rate": 4.9961826771182784e-05,
"loss": 0.6086,
"num_input_tokens_seen": 768056,
"step": 19
},
{
"epoch": 0.3669724770642202,
"grad_norm": 0.06627275049686432,
"learning_rate": 4.995770395678171e-05,
"loss": 0.4591,
"num_input_tokens_seen": 806256,
"step": 20
},
{
"epoch": 0.3853211009174312,
"grad_norm": 0.05830290913581848,
"learning_rate": 4.9953369960546676e-05,
"loss": 0.3731,
"num_input_tokens_seen": 842336,
"step": 21
},
{
"epoch": 0.4036697247706422,
"grad_norm": 0.07277437299489975,
"learning_rate": 4.9948824819150185e-05,
"loss": 0.6243,
"num_input_tokens_seen": 876672,
"step": 22
},
{
"epoch": 0.42201834862385323,
"grad_norm": 0.07477546483278275,
"learning_rate": 4.994406857105136e-05,
"loss": 0.5788,
"num_input_tokens_seen": 915192,
"step": 23
},
{
"epoch": 0.44036697247706424,
"grad_norm": 0.06912907212972641,
"learning_rate": 4.993910125649561e-05,
"loss": 0.4753,
"num_input_tokens_seen": 951904,
"step": 24
},
{
"epoch": 0.45871559633027525,
"grad_norm": 0.0655476376414299,
"learning_rate": 4.993392291751431e-05,
"loss": 0.4518,
"num_input_tokens_seen": 1001816,
"step": 25
},
{
"epoch": 0.47706422018348627,
"grad_norm": 0.06466512382030487,
"learning_rate": 4.992853359792444e-05,
"loss": 0.5638,
"num_input_tokens_seen": 1053064,
"step": 26
},
{
"epoch": 0.4954128440366973,
"grad_norm": 0.0645688995718956,
"learning_rate": 4.99229333433282e-05,
"loss": 0.4644,
"num_input_tokens_seen": 1086688,
"step": 27
},
{
"epoch": 0.5137614678899083,
"grad_norm": 0.07181251049041748,
"learning_rate": 4.9917122201112656e-05,
"loss": 0.6191,
"num_input_tokens_seen": 1134824,
"step": 28
},
{
"epoch": 0.5321100917431193,
"grad_norm": 0.07322589308023453,
"learning_rate": 4.9911100220449293e-05,
"loss": 0.6752,
"num_input_tokens_seen": 1172072,
"step": 29
},
{
"epoch": 0.5504587155963303,
"grad_norm": 0.06396070122718811,
"learning_rate": 4.990486745229364e-05,
"loss": 0.3587,
"num_input_tokens_seen": 1211096,
"step": 30
},
{
"epoch": 0.5688073394495413,
"grad_norm": 0.07803395390510559,
"learning_rate": 4.989842394938482e-05,
"loss": 0.459,
"num_input_tokens_seen": 1259456,
"step": 31
},
{
"epoch": 0.5871559633027523,
"grad_norm": 0.05974648892879486,
"learning_rate": 4.989176976624511e-05,
"loss": 0.4148,
"num_input_tokens_seen": 1306944,
"step": 32
},
{
"epoch": 0.6055045871559633,
"grad_norm": 0.09784268587827682,
"learning_rate": 4.988490495917947e-05,
"loss": 0.539,
"num_input_tokens_seen": 1353744,
"step": 33
},
{
"epoch": 0.6238532110091743,
"grad_norm": 0.09906516224145889,
"learning_rate": 4.987782958627508e-05,
"loss": 0.5453,
"num_input_tokens_seen": 1394736,
"step": 34
},
{
"epoch": 0.6422018348623854,
"grad_norm": 0.08984062820672989,
"learning_rate": 4.987054370740083e-05,
"loss": 0.468,
"num_input_tokens_seen": 1442048,
"step": 35
},
{
"epoch": 0.6605504587155964,
"grad_norm": 0.08672655373811722,
"learning_rate": 4.9863047384206835e-05,
"loss": 0.4078,
"num_input_tokens_seen": 1478440,
"step": 36
},
{
"epoch": 0.6788990825688074,
"grad_norm": 0.1327345073223114,
"learning_rate": 4.9855340680123905e-05,
"loss": 0.5299,
"num_input_tokens_seen": 1525992,
"step": 37
},
{
"epoch": 0.6972477064220184,
"grad_norm": 0.09178602695465088,
"learning_rate": 4.9847423660363e-05,
"loss": 0.439,
"num_input_tokens_seen": 1555608,
"step": 38
},
{
"epoch": 0.7155963302752294,
"grad_norm": 0.09418320655822754,
"learning_rate": 4.983929639191469e-05,
"loss": 0.5337,
"num_input_tokens_seen": 1597392,
"step": 39
},
{
"epoch": 0.7339449541284404,
"grad_norm": 0.08294719457626343,
"learning_rate": 4.983095894354858e-05,
"loss": 0.4536,
"num_input_tokens_seen": 1649656,
"step": 40
},
{
"epoch": 0.7522935779816514,
"grad_norm": 0.09774205833673477,
"learning_rate": 4.982241138581273e-05,
"loss": 0.5221,
"num_input_tokens_seen": 1695952,
"step": 41
},
{
"epoch": 0.7706422018348624,
"grad_norm": 0.09319107979536057,
"learning_rate": 4.9813653791033057e-05,
"loss": 0.4279,
"num_input_tokens_seen": 1737224,
"step": 42
},
{
"epoch": 0.7889908256880734,
"grad_norm": 0.09561405330896378,
"learning_rate": 4.980468623331273e-05,
"loss": 0.5121,
"num_input_tokens_seen": 1772320,
"step": 43
},
{
"epoch": 0.8073394495412844,
"grad_norm": 0.08274025470018387,
"learning_rate": 4.979550878853154e-05,
"loss": 0.54,
"num_input_tokens_seen": 1823888,
"step": 44
},
{
"epoch": 0.8256880733944955,
"grad_norm": 0.08728913217782974,
"learning_rate": 4.9786121534345265e-05,
"loss": 0.4488,
"num_input_tokens_seen": 1872488,
"step": 45
},
{
"epoch": 0.8440366972477065,
"grad_norm": 0.0787016749382019,
"learning_rate": 4.9776524550184965e-05,
"loss": 0.4353,
"num_input_tokens_seen": 1924744,
"step": 46
},
{
"epoch": 0.8623853211009175,
"grad_norm": 0.10952188074588776,
"learning_rate": 4.97667179172564e-05,
"loss": 0.4784,
"num_input_tokens_seen": 1959936,
"step": 47
},
{
"epoch": 0.8807339449541285,
"grad_norm": 0.08525826781988144,
"learning_rate": 4.975670171853926e-05,
"loss": 0.3586,
"num_input_tokens_seen": 2003896,
"step": 48
},
{
"epoch": 0.8990825688073395,
"grad_norm": 0.10409987717866898,
"learning_rate": 4.9746476038786496e-05,
"loss": 0.4451,
"num_input_tokens_seen": 2047632,
"step": 49
},
{
"epoch": 0.9174311926605505,
"grad_norm": 0.0782993957400322,
"learning_rate": 4.973604096452361e-05,
"loss": 0.3591,
"num_input_tokens_seen": 2096928,
"step": 50
},
{
"epoch": 0.9357798165137615,
"grad_norm": 0.09829951077699661,
"learning_rate": 4.9725396584047925e-05,
"loss": 0.3415,
"num_input_tokens_seen": 2129536,
"step": 51
},
{
"epoch": 0.9541284403669725,
"grad_norm": 0.10606162995100021,
"learning_rate": 4.971454298742779e-05,
"loss": 0.3758,
"num_input_tokens_seen": 2169144,
"step": 52
},
{
"epoch": 0.9724770642201835,
"grad_norm": 0.09280356764793396,
"learning_rate": 4.97034802665019e-05,
"loss": 0.485,
"num_input_tokens_seen": 2207720,
"step": 53
},
{
"epoch": 0.9908256880733946,
"grad_norm": 0.11888203024864197,
"learning_rate": 4.9692208514878444e-05,
"loss": 0.3469,
"num_input_tokens_seen": 2236392,
"step": 54
}
],
"logging_steps": 1.0,
"max_steps": 1080,
"num_input_tokens_seen": 2236392,
"num_train_epochs": 20,
"save_steps": 54,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.8866577009855693e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}