lora / trainer_state.json
valoomba's picture
Upload folder using huggingface_hub
d72b603 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.864,
"eval_steps": 500,
"global_step": 155,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.16,
"grad_norm": 0.04806969687342644,
"learning_rate": 6.249999999999999e-07,
"loss": 1.1372,
"mean_token_accuracy": 0.7175553351640701,
"step": 5
},
{
"epoch": 0.32,
"grad_norm": 0.04902997612953186,
"learning_rate": 9.99543333708549e-07,
"loss": 1.1327,
"mean_token_accuracy": 0.7127479765564203,
"step": 10
},
{
"epoch": 0.48,
"grad_norm": 0.052965614944696426,
"learning_rate": 9.944154131125642e-07,
"loss": 1.1403,
"mean_token_accuracy": 0.7101998340338469,
"step": 15
},
{
"epoch": 0.64,
"grad_norm": 0.049492210149765015,
"learning_rate": 9.836474315195147e-07,
"loss": 1.1492,
"mean_token_accuracy": 0.7093485150486231,
"step": 20
},
{
"epoch": 0.8,
"grad_norm": 0.04770048335194588,
"learning_rate": 9.673622250534155e-07,
"loss": 1.0896,
"mean_token_accuracy": 0.722656001150608,
"step": 25
},
{
"epoch": 0.96,
"grad_norm": 0.04960792139172554,
"learning_rate": 9.457455677726447e-07,
"loss": 1.1031,
"mean_token_accuracy": 0.7249287318438291,
"step": 30
},
{
"epoch": 1.096,
"grad_norm": 0.05211903899908066,
"learning_rate": 9.190440524459202e-07,
"loss": 1.0969,
"mean_token_accuracy": 0.7173009343883571,
"step": 35
},
{
"epoch": 1.256,
"grad_norm": 0.05225418508052826,
"learning_rate": 8.875622775367259e-07,
"loss": 1.1024,
"mean_token_accuracy": 0.7184804223477841,
"step": 40
},
{
"epoch": 1.416,
"grad_norm": 0.053441036492586136,
"learning_rate": 8.516593724857597e-07,
"loss": 1.1062,
"mean_token_accuracy": 0.7180147383362054,
"step": 45
},
{
"epoch": 1.576,
"grad_norm": 0.05642622709274292,
"learning_rate": 8.117449009293668e-07,
"loss": 1.1143,
"mean_token_accuracy": 0.7181682541966439,
"step": 50
},
{
"epoch": 1.736,
"grad_norm": 0.06203881651163101,
"learning_rate": 7.682741885881314e-07,
"loss": 1.1619,
"mean_token_accuracy": 0.7082567475736141,
"step": 55
},
{
"epoch": 1.896,
"grad_norm": 0.05656283721327782,
"learning_rate": 7.217431291229067e-07,
"loss": 1.1153,
"mean_token_accuracy": 0.722714689001441,
"step": 60
},
{
"epoch": 2.032,
"grad_norm": 0.061042506247758865,
"learning_rate": 6.726825272106538e-07,
"loss": 1.1493,
"mean_token_accuracy": 0.7165276228505022,
"step": 65
},
{
"epoch": 2.192,
"grad_norm": 0.058021657168865204,
"learning_rate": 6.216520433716544e-07,
"loss": 1.1358,
"mean_token_accuracy": 0.7132049109786749,
"step": 70
},
{
"epoch": 2.352,
"grad_norm": 0.059838637709617615,
"learning_rate": 5.69233809622687e-07,
"loss": 1.1148,
"mean_token_accuracy": 0.7178624272346497,
"step": 75
},
{
"epoch": 2.512,
"grad_norm": 0.06404729187488556,
"learning_rate": 5.160257887858277e-07,
"loss": 1.1569,
"mean_token_accuracy": 0.7101449474692345,
"step": 80
},
{
"epoch": 2.672,
"grad_norm": 0.06028318777680397,
"learning_rate": 4.626349532067879e-07,
"loss": 1.0867,
"mean_token_accuracy": 0.7239396564662457,
"step": 85
},
{
"epoch": 2.832,
"grad_norm": 0.06384854018688202,
"learning_rate": 4.096703606968006e-07,
"loss": 1.1188,
"mean_token_accuracy": 0.7177162211388349,
"step": 90
},
{
"epoch": 2.992,
"grad_norm": 0.06171684339642525,
"learning_rate": 3.577362066844838e-07,
"loss": 1.1188,
"mean_token_accuracy": 0.7161729197949172,
"step": 95
},
{
"epoch": 3.128,
"grad_norm": 0.0635930597782135,
"learning_rate": 3.0742493183550454e-07,
"loss": 1.0999,
"mean_token_accuracy": 0.714051765992361,
"step": 100
},
{
"epoch": 3.288,
"grad_norm": 0.06290465593338013,
"learning_rate": 2.593104637651087e-07,
"loss": 1.138,
"mean_token_accuracy": 0.7161103874444962,
"step": 105
},
{
"epoch": 3.448,
"grad_norm": 0.06601794064044952,
"learning_rate": 2.1394166993891526e-07,
"loss": 1.1073,
"mean_token_accuracy": 0.7206942658871412,
"step": 110
},
{
"epoch": 3.608,
"grad_norm": 0.06451869755983353,
"learning_rate": 1.7183609644824092e-07,
"loss": 1.105,
"mean_token_accuracy": 0.7173311490565538,
"step": 115
},
{
"epoch": 3.768,
"grad_norm": 0.06503637880086899,
"learning_rate": 1.3347406408508694e-07,
"loss": 1.1026,
"mean_token_accuracy": 0.7226938724517822,
"step": 120
},
{
"epoch": 3.928,
"grad_norm": 0.06369519233703613,
"learning_rate": 9.929318906602174e-08,
"loss": 1.136,
"mean_token_accuracy": 0.7119275834411383,
"step": 125
},
{
"epoch": 4.064,
"grad_norm": 0.0652015432715416,
"learning_rate": 6.968339090999186e-08,
"loss": 1.1041,
"mean_token_accuracy": 0.7163285718244665,
"step": 130
},
{
"epoch": 4.224,
"grad_norm": 0.06359368562698364,
"learning_rate": 4.498244441786675e-08,
"loss": 1.1359,
"mean_token_accuracy": 0.7156530544161797,
"step": 135
},
{
"epoch": 4.384,
"grad_norm": 0.06807275116443634,
"learning_rate": 2.547212649466568e-08,
"loss": 1.1229,
"mean_token_accuracy": 0.7130052808672189,
"step": 140
},
{
"epoch": 4.5440000000000005,
"grad_norm": 0.06558381021022797,
"learning_rate": 1.1375001769727999e-08,
"loss": 1.0983,
"mean_token_accuracy": 0.7218218572437763,
"step": 145
},
{
"epoch": 4.704,
"grad_norm": 0.06619936972856522,
"learning_rate": 2.851883682973233e-09,
"loss": 1.1149,
"mean_token_accuracy": 0.7202201712876558,
"step": 150
},
{
"epoch": 4.864,
"grad_norm": 0.0672445297241211,
"learning_rate": 0.0,
"loss": 1.0885,
"mean_token_accuracy": 0.7229270905256271,
"step": 155
}
],
"logging_steps": 5,
"max_steps": 155,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 9.40617510529582e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}