gemma-3-12b-it-sft / trainer_state.json
pvbhanuteja's picture
Model save
fcb9999 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9998447060285119,
"eval_steps": 500,
"global_step": 503,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.019877628350467434,
"grad_norm": 49.489271990521864,
"learning_rate": 1.730769230769231e-05,
"loss": 18.2674,
"num_tokens": 35435900.0,
"step": 10
},
{
"epoch": 0.03975525670093487,
"grad_norm": 94.13839060488317,
"learning_rate": 3.653846153846154e-05,
"loss": 6.0352,
"num_tokens": 70986000.0,
"step": 20
},
{
"epoch": 0.0596328850514023,
"grad_norm": 36.22989263913226,
"learning_rate": 4.999560818649105e-05,
"loss": 2.257,
"num_tokens": 106406261.0,
"step": 30
},
{
"epoch": 0.07951051340186974,
"grad_norm": 12.381662510455556,
"learning_rate": 4.99175791880267e-05,
"loss": 1.6806,
"num_tokens": 142007549.0,
"step": 40
},
{
"epoch": 0.09938814175233718,
"grad_norm": 6.078747336589003,
"learning_rate": 4.9742343823849924e-05,
"loss": 1.6283,
"num_tokens": 177749958.0,
"step": 50
},
{
"epoch": 0.1192657701028046,
"grad_norm": 45.316378371707856,
"learning_rate": 4.947066194375331e-05,
"loss": 1.405,
"num_tokens": 213223619.0,
"step": 60
},
{
"epoch": 0.13914339845327203,
"grad_norm": 4.457640477421176,
"learning_rate": 4.910371160572305e-05,
"loss": 1.4731,
"num_tokens": 248699816.0,
"step": 70
},
{
"epoch": 0.15902102680373947,
"grad_norm": 3.8571852421472106,
"learning_rate": 4.864308396768294e-05,
"loss": 1.1962,
"num_tokens": 284242373.0,
"step": 80
},
{
"epoch": 0.17889865515420691,
"grad_norm": 2.922271088762262,
"learning_rate": 4.8090776387967915e-05,
"loss": 1.1482,
"num_tokens": 319801798.0,
"step": 90
},
{
"epoch": 0.19877628350467436,
"grad_norm": 3.07929680688637,
"learning_rate": 4.7449183764444924e-05,
"loss": 1.1573,
"num_tokens": 355227417.0,
"step": 100
},
{
"epoch": 0.2186539118551418,
"grad_norm": 2.986142143054458,
"learning_rate": 4.672108814983597e-05,
"loss": 1.1133,
"num_tokens": 35532164.0,
"step": 110
},
{
"epoch": 0.2385315402056092,
"grad_norm": 2.347646077232914,
"learning_rate": 4.590964668827306e-05,
"loss": 1.1624,
"num_tokens": 71080907.0,
"step": 120
},
{
"epoch": 0.25840916855607665,
"grad_norm": 3.077510825746025,
"learning_rate": 4.501837792539419e-05,
"loss": 1.0658,
"num_tokens": 106602450.0,
"step": 130
},
{
"epoch": 0.27828679690654406,
"grad_norm": 1.7347828183409042,
"learning_rate": 4.4051146551342136e-05,
"loss": 1.0483,
"num_tokens": 142169486.0,
"step": 140
},
{
"epoch": 0.29816442525701153,
"grad_norm": 2.802057431831608,
"learning_rate": 4.301214664282277e-05,
"loss": 1.0333,
"num_tokens": 177675215.0,
"step": 150
},
{
"epoch": 0.31804205360747895,
"grad_norm": 2.268188963759059,
"learning_rate": 4.190588347688833e-05,
"loss": 1.0238,
"num_tokens": 213190309.0,
"step": 160
},
{
"epoch": 0.3379196819579464,
"grad_norm": 1.6440471046115712,
"learning_rate": 4.073715399530405e-05,
"loss": 0.9957,
"num_tokens": 248802333.0,
"step": 170
},
{
"epoch": 0.35779731030841383,
"grad_norm": 1.3717651355528624,
"learning_rate": 3.9511026004207795e-05,
"loss": 0.9574,
"num_tokens": 284367601.0,
"step": 180
},
{
"epoch": 0.37767493865888124,
"grad_norm": 1.6711348270758677,
"learning_rate": 3.823281619925673e-05,
"loss": 0.9611,
"num_tokens": 319996464.0,
"step": 190
},
{
"epoch": 0.3975525670093487,
"grad_norm": 1.8188583401200056,
"learning_rate": 3.69080671115473e-05,
"loss": 0.9318,
"num_tokens": 355585930.0,
"step": 200
},
{
"epoch": 0.4174301953598161,
"grad_norm": 1.6314158389343252,
"learning_rate": 3.554252307427526e-05,
"loss": 0.9428,
"num_tokens": 391200376.0,
"step": 210
},
{
"epoch": 0.4373078237102836,
"grad_norm": 1.406291241468206,
"learning_rate": 3.414210531434781e-05,
"loss": 0.934,
"num_tokens": 426715690.0,
"step": 220
},
{
"epoch": 0.457185452060751,
"grad_norm": 1.5568298850166404,
"learning_rate": 3.271288627695541e-05,
"loss": 0.8906,
"num_tokens": 462193838.0,
"step": 230
},
{
"epoch": 0.4770630804112184,
"grad_norm": 1.6374375121178215,
"learning_rate": 3.126106329443562e-05,
"loss": 0.8869,
"num_tokens": 497775413.0,
"step": 240
},
{
"epoch": 0.4969407087616859,
"grad_norm": 1.4278396281710222,
"learning_rate": 2.9792931713605558e-05,
"loss": 0.8715,
"num_tokens": 533288808.0,
"step": 250
},
{
"epoch": 0.5168183371121533,
"grad_norm": 1.3315422073102914,
"learning_rate": 2.831485759808712e-05,
"loss": 0.8535,
"num_tokens": 568733244.0,
"step": 260
},
{
"epoch": 0.5366959654626208,
"grad_norm": 1.4292589501021156,
"learning_rate": 2.683325012399241e-05,
"loss": 0.8555,
"num_tokens": 604337940.0,
"step": 270
},
{
"epoch": 0.5565735938130881,
"grad_norm": 1.3191036037659456,
"learning_rate": 2.5354533788666173e-05,
"loss": 0.8258,
"num_tokens": 639897455.0,
"step": 280
},
{
"epoch": 0.5764512221635556,
"grad_norm": 1.2135834698476922,
"learning_rate": 2.388512055299283e-05,
"loss": 0.8152,
"num_tokens": 675560731.0,
"step": 290
},
{
"epoch": 0.5963288505140231,
"grad_norm": 1.1849646162622842,
"learning_rate": 2.2431382038063615e-05,
"loss": 0.8033,
"num_tokens": 710895149.0,
"step": 300
},
{
"epoch": 0.6162064788644904,
"grad_norm": 1.0526968604787708,
"learning_rate": 2.0999621896763955e-05,
"loss": 0.7823,
"num_tokens": 746372135.0,
"step": 310
},
{
"epoch": 0.6360841072149579,
"grad_norm": 1.0814389110808431,
"learning_rate": 1.959604848008217e-05,
"loss": 0.7653,
"num_tokens": 781964693.0,
"step": 320
},
{
"epoch": 0.6559617355654254,
"grad_norm": 1.1579079388621223,
"learning_rate": 1.8226747916663184e-05,
"loss": 0.7535,
"num_tokens": 817483499.0,
"step": 330
},
{
"epoch": 0.6758393639158928,
"grad_norm": 1.0764253018415963,
"learning_rate": 1.6897657722338715e-05,
"loss": 0.7417,
"num_tokens": 853175363.0,
"step": 340
},
{
"epoch": 0.6957169922663602,
"grad_norm": 1.0645270530767825,
"learning_rate": 1.561454105406746e-05,
"loss": 0.7284,
"num_tokens": 888676759.0,
"step": 350
},
{
"epoch": 0.7155946206168277,
"grad_norm": 1.0991349699273438,
"learning_rate": 1.4382961719924382e-05,
"loss": 0.7102,
"num_tokens": 924126537.0,
"step": 360
},
{
"epoch": 0.7354722489672951,
"grad_norm": 1.096361260557103,
"learning_rate": 1.3208260053500118e-05,
"loss": 0.7016,
"num_tokens": 959627625.0,
"step": 370
},
{
"epoch": 0.7553498773177625,
"grad_norm": 1.2001309900905168,
"learning_rate": 1.2095529757323018e-05,
"loss": 0.6735,
"num_tokens": 995276803.0,
"step": 380
},
{
"epoch": 0.77522750566823,
"grad_norm": 1.0243664466225824,
"learning_rate": 1.104959581571463e-05,
"loss": 0.6649,
"num_tokens": 1030735793.0,
"step": 390
},
{
"epoch": 0.7951051340186974,
"grad_norm": 1.4526878564718964,
"learning_rate": 1.0074993572852243e-05,
"loss": 0.6713,
"num_tokens": 1066374499.0,
"step": 400
},
{
"epoch": 0.8149827623691648,
"grad_norm": 1.050605567010305,
"learning_rate": 9.175949066759368e-06,
"loss": 0.6419,
"num_tokens": 1101855303.0,
"step": 410
},
{
"epoch": 0.8348603907196323,
"grad_norm": 1.125778801254232,
"learning_rate": 8.356360704499399e-06,
"loss": 0.635,
"num_tokens": 1137354673.0,
"step": 420
},
{
"epoch": 0.8547380190700997,
"grad_norm": 1.1832572011409386,
"learning_rate": 7.619782358031821e-06,
"loss": 0.6267,
"num_tokens": 1172860263.0,
"step": 430
},
{
"epoch": 0.8746156474205672,
"grad_norm": 1.1054714685819218,
"learning_rate": 6.969407954030057e-06,
"loss": 0.6116,
"num_tokens": 1208508537.0,
"step": 440
},
{
"epoch": 0.8944932757710345,
"grad_norm": 1.089754156750933,
"learning_rate": 6.40805762448228e-06,
"loss": 0.6051,
"num_tokens": 1244052723.0,
"step": 450
},
{
"epoch": 0.914370904121502,
"grad_norm": 1.0612130068358916,
"learning_rate": 5.938165478128298e-06,
"loss": 0.591,
"num_tokens": 1279456435.0,
"step": 460
},
{
"epoch": 0.9342485324719695,
"grad_norm": 1.1570486650093461,
"learning_rate": 5.5617690457577016e-06,
"loss": 0.5808,
"num_tokens": 1314958319.0,
"step": 470
},
{
"epoch": 0.9541261608224368,
"grad_norm": 1.4354838420805704,
"learning_rate": 5.280500445136247e-06,
"loss": 0.5754,
"num_tokens": 1350578878.0,
"step": 480
},
{
"epoch": 0.9740037891729043,
"grad_norm": 1.3310227594903838,
"learning_rate": 5.095579303870732e-06,
"loss": 0.5693,
"num_tokens": 1386062688.0,
"step": 490
},
{
"epoch": 0.9938814175233718,
"grad_norm": 1.0908684710720695,
"learning_rate": 5.007807470900112e-06,
"loss": 0.5606,
"num_tokens": 1421565738.0,
"step": 500
},
{
"epoch": 0.9998447060285119,
"num_tokens": 1432289557.0,
"step": 503,
"total_flos": 1.1360204825667672e+20,
"train_loss": 0.6426609048312512,
"train_runtime": 119967.1502,
"train_samples_per_second": 2.147,
"train_steps_per_second": 0.004
}
],
"logging_steps": 10,
"max_steps": 503,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.1360204825667672e+20,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}