|
{ |
|
"best_metric": 0.2546990215778351, |
|
"best_model_checkpoint": "xlm-roberta-base-all-finetuned-toxicity-classification/checkpoint-2532", |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 7596, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8683517640863615e-05, |
|
"loss": 0.2799, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.7367035281727228e-05, |
|
"loss": 0.2876, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.6050552922590838e-05, |
|
"loss": 0.2671, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.473407056345445e-05, |
|
"loss": 0.2639, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3417588204318064e-05, |
|
"loss": 0.2675, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.2546990215778351, |
|
"eval_runtime": 35.0277, |
|
"eval_samples_per_second": 128.47, |
|
"eval_steps_per_second": 8.051, |
|
"step": 2532 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2101105845181676e-05, |
|
"loss": 0.2072, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.0784623486045287e-05, |
|
"loss": 0.2035, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.4681411269089e-06, |
|
"loss": 0.1987, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.151658767772512e-06, |
|
"loss": 0.1938, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 6.835176408636125e-06, |
|
"loss": 0.189, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.29669827222824097, |
|
"eval_runtime": 35.0287, |
|
"eval_samples_per_second": 128.466, |
|
"eval_steps_per_second": 8.051, |
|
"step": 5064 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 5.518694049499738e-06, |
|
"loss": 0.1412, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.20221169036335e-06, |
|
"loss": 0.1467, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.885729331226962e-06, |
|
"loss": 0.1397, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.5692469720905742e-06, |
|
"loss": 0.1476, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.527646129541864e-07, |
|
"loss": 0.1383, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.4095066785812378, |
|
"eval_runtime": 35.0202, |
|
"eval_samples_per_second": 128.497, |
|
"eval_steps_per_second": 8.053, |
|
"step": 7596 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 7596, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"total_flos": 3.196799322624e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|