luganda-nllb-qlora-checkpoints / trainer_state.json
kserumaga's picture
Graceful shutdown
b8b5684 verified
{
"best_global_step": 2000,
"best_metric": 0.0787040963768959,
"best_model_checkpoint": "/tmp/results/checkpoint-2000",
"epoch": 2.220324979344533,
"eval_steps": 200,
"global_step": 2015,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.055081244836133296,
"grad_norm": 0.15732981264591217,
"learning_rate": 0.00019640234948604994,
"loss": 1.2703,
"step": 50
},
{
"epoch": 0.11016248967226659,
"grad_norm": 0.13457506895065308,
"learning_rate": 0.00019273127753303965,
"loss": 0.2101,
"step": 100
},
{
"epoch": 0.16524373450839988,
"grad_norm": 0.11798923462629318,
"learning_rate": 0.00018906020558002938,
"loss": 0.1801,
"step": 150
},
{
"epoch": 0.22032497934453318,
"grad_norm": 0.10562872886657715,
"learning_rate": 0.00018538913362701911,
"loss": 0.1621,
"step": 200
},
{
"epoch": 0.22032497934453318,
"eval_loss": 0.15831586718559265,
"eval_runtime": 95.9669,
"eval_samples_per_second": 33.626,
"eval_steps_per_second": 4.21,
"step": 200
},
{
"epoch": 0.27540622418066646,
"grad_norm": 0.11607076972723007,
"learning_rate": 0.00018171806167400882,
"loss": 0.1526,
"step": 250
},
{
"epoch": 0.33048746901679976,
"grad_norm": 0.20103491842746735,
"learning_rate": 0.00017804698972099853,
"loss": 0.1498,
"step": 300
},
{
"epoch": 0.38556871385293306,
"grad_norm": 0.20092593133449554,
"learning_rate": 0.00017437591776798826,
"loss": 0.1481,
"step": 350
},
{
"epoch": 0.44064995868906637,
"grad_norm": 0.15526321530342102,
"learning_rate": 0.000170704845814978,
"loss": 0.1533,
"step": 400
},
{
"epoch": 0.44064995868906637,
"eval_loss": 0.13016295433044434,
"eval_runtime": 96.1932,
"eval_samples_per_second": 33.547,
"eval_steps_per_second": 4.2,
"step": 400
},
{
"epoch": 0.49573120352519967,
"grad_norm": 0.23773688077926636,
"learning_rate": 0.0001670337738619677,
"loss": 0.1434,
"step": 450
},
{
"epoch": 0.5508124483613329,
"grad_norm": 0.14756543934345245,
"learning_rate": 0.00016336270190895743,
"loss": 0.1316,
"step": 500
},
{
"epoch": 0.6058936931974662,
"grad_norm": 0.19601435959339142,
"learning_rate": 0.00015969162995594716,
"loss": 0.1301,
"step": 550
},
{
"epoch": 0.6609749380335995,
"grad_norm": 0.1771874874830246,
"learning_rate": 0.00015602055800293687,
"loss": 0.1215,
"step": 600
},
{
"epoch": 0.6609749380335995,
"eval_loss": 0.11715047806501389,
"eval_runtime": 95.8683,
"eval_samples_per_second": 33.661,
"eval_steps_per_second": 4.214,
"step": 600
},
{
"epoch": 0.7160561828697328,
"grad_norm": 0.16557306051254272,
"learning_rate": 0.00015234948604992657,
"loss": 0.1139,
"step": 650
},
{
"epoch": 0.7711374277058661,
"grad_norm": 0.2927381992340088,
"learning_rate": 0.0001486784140969163,
"loss": 0.1214,
"step": 700
},
{
"epoch": 0.8262186725419994,
"grad_norm": 0.3338780701160431,
"learning_rate": 0.00014500734214390604,
"loss": 0.1285,
"step": 750
},
{
"epoch": 0.8812999173781327,
"grad_norm": 0.2900579273700714,
"learning_rate": 0.00014133627019089574,
"loss": 0.1274,
"step": 800
},
{
"epoch": 0.8812999173781327,
"eval_loss": 0.10116679221391678,
"eval_runtime": 94.9307,
"eval_samples_per_second": 33.993,
"eval_steps_per_second": 4.256,
"step": 800
},
{
"epoch": 0.936381162214266,
"grad_norm": 0.2934402823448181,
"learning_rate": 0.00013766519823788548,
"loss": 0.1169,
"step": 850
},
{
"epoch": 0.9914624070503993,
"grad_norm": 0.2068411260843277,
"learning_rate": 0.0001339941262848752,
"loss": 0.1179,
"step": 900
},
{
"epoch": 1.0473698705590746,
"grad_norm": 0.10344358533620834,
"learning_rate": 0.0001303230543318649,
"loss": 0.1003,
"step": 950
},
{
"epoch": 1.102451115395208,
"grad_norm": 0.24283887445926666,
"learning_rate": 0.00012665198237885462,
"loss": 0.1125,
"step": 1000
},
{
"epoch": 1.102451115395208,
"eval_loss": 0.09346170723438263,
"eval_runtime": 98.7546,
"eval_samples_per_second": 32.677,
"eval_steps_per_second": 4.091,
"step": 1000
},
{
"epoch": 1.1575323602313412,
"grad_norm": 0.14652925729751587,
"learning_rate": 0.00012298091042584435,
"loss": 0.1058,
"step": 1050
},
{
"epoch": 1.2126136050674745,
"grad_norm": 0.23155081272125244,
"learning_rate": 0.00011930983847283408,
"loss": 0.114,
"step": 1100
},
{
"epoch": 1.2676948499036078,
"grad_norm": 0.22973254323005676,
"learning_rate": 0.00011563876651982379,
"loss": 0.1065,
"step": 1150
},
{
"epoch": 1.3227760947397411,
"grad_norm": 0.35016024112701416,
"learning_rate": 0.00011196769456681351,
"loss": 0.117,
"step": 1200
},
{
"epoch": 1.3227760947397411,
"eval_loss": 0.0892082005739212,
"eval_runtime": 99.5776,
"eval_samples_per_second": 32.407,
"eval_steps_per_second": 4.057,
"step": 1200
},
{
"epoch": 1.3767557146791518,
"grad_norm": 0.2444484978914261,
"learning_rate": 0.00010829662261380324,
"loss": 0.1042,
"step": 1250
},
{
"epoch": 1.4318369595152851,
"grad_norm": 0.21167899668216705,
"learning_rate": 0.00010462555066079296,
"loss": 0.1111,
"step": 1300
},
{
"epoch": 1.4869182043514184,
"grad_norm": 0.282637357711792,
"learning_rate": 0.00010095447870778268,
"loss": 0.0953,
"step": 1350
},
{
"epoch": 1.5419994491875517,
"grad_norm": 0.159465953707695,
"learning_rate": 9.72834067547724e-05,
"loss": 0.1015,
"step": 1400
},
{
"epoch": 1.5419994491875517,
"eval_loss": 0.08904121816158295,
"eval_runtime": 98.1262,
"eval_samples_per_second": 32.886,
"eval_steps_per_second": 4.117,
"step": 1400
},
{
"epoch": 1.5970806940236848,
"grad_norm": 0.30280402302742004,
"learning_rate": 9.361233480176212e-05,
"loss": 0.0955,
"step": 1450
},
{
"epoch": 1.6521619388598183,
"grad_norm": 0.2655673623085022,
"learning_rate": 8.994126284875184e-05,
"loss": 0.0975,
"step": 1500
},
{
"epoch": 1.7072431836959514,
"grad_norm": 0.3565061688423157,
"learning_rate": 8.627019089574157e-05,
"loss": 0.081,
"step": 1550
},
{
"epoch": 1.762324428532085,
"grad_norm": 0.5034757852554321,
"learning_rate": 8.259911894273127e-05,
"loss": 0.1027,
"step": 1600
},
{
"epoch": 1.762324428532085,
"eval_loss": 0.08446714282035828,
"eval_runtime": 98.0553,
"eval_samples_per_second": 32.91,
"eval_steps_per_second": 4.12,
"step": 1600
},
{
"epoch": 1.817405673368218,
"grad_norm": 0.20851312577724457,
"learning_rate": 7.892804698972101e-05,
"loss": 0.0979,
"step": 1650
},
{
"epoch": 1.8724869182043515,
"grad_norm": 0.28781506419181824,
"learning_rate": 7.525697503671073e-05,
"loss": 0.1001,
"step": 1700
},
{
"epoch": 1.9275681630404846,
"grad_norm": 0.2584012448787689,
"learning_rate": 7.158590308370045e-05,
"loss": 0.0978,
"step": 1750
},
{
"epoch": 1.9826494078766181,
"grad_norm": 0.2780962288379669,
"learning_rate": 6.791483113069016e-05,
"loss": 0.0976,
"step": 1800
},
{
"epoch": 1.9826494078766181,
"eval_loss": 0.08305753767490387,
"eval_runtime": 98.1323,
"eval_samples_per_second": 32.884,
"eval_steps_per_second": 4.117,
"step": 1800
},
{
"epoch": 2.0385568713852935,
"grad_norm": 0.15813182294368744,
"learning_rate": 6.424375917767988e-05,
"loss": 0.089,
"step": 1850
},
{
"epoch": 2.0936381162214266,
"grad_norm": 0.2571295499801636,
"learning_rate": 6.057268722466961e-05,
"loss": 0.0912,
"step": 1900
},
{
"epoch": 2.1487193610575597,
"grad_norm": 0.27553290128707886,
"learning_rate": 5.690161527165933e-05,
"loss": 0.0922,
"step": 1950
},
{
"epoch": 2.203800605893693,
"grad_norm": 0.2373432070016861,
"learning_rate": 5.3230543318649054e-05,
"loss": 0.0963,
"step": 2000
},
{
"epoch": 2.203800605893693,
"eval_loss": 0.0787040963768959,
"eval_runtime": 96.6596,
"eval_samples_per_second": 33.385,
"eval_steps_per_second": 4.18,
"step": 2000
}
],
"logging_steps": 50,
"max_steps": 2724,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 411986682593280.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}