ThinkGuard / trainer_state.json
Rakancorle1's picture
End of training
0ff0063 verified
raw
history blame
23 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.994262174488745,
"eval_steps": 500,
"global_step": 1272,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02353979696925114,
"grad_norm": 8.782275747399924,
"learning_rate": 3.846153846153846e-06,
"loss": 1.5306,
"step": 10
},
{
"epoch": 0.04707959393850228,
"grad_norm": 4.3210560861300085,
"learning_rate": 7.692307692307692e-06,
"loss": 0.8302,
"step": 20
},
{
"epoch": 0.07061939090775342,
"grad_norm": 3.023277309045724,
"learning_rate": 1.153846153846154e-05,
"loss": 0.6217,
"step": 30
},
{
"epoch": 0.09415918787700456,
"grad_norm": 2.654024929474284,
"learning_rate": 1.4999975655288908e-05,
"loss": 0.5967,
"step": 40
},
{
"epoch": 0.1176989848462557,
"grad_norm": 2.8603296276880723,
"learning_rate": 1.4997054481186253e-05,
"loss": 0.5732,
"step": 50
},
{
"epoch": 0.14123878181550684,
"grad_norm": 2.4252498205382347,
"learning_rate": 1.4989266537738068e-05,
"loss": 0.5573,
"step": 60
},
{
"epoch": 0.164778578784758,
"grad_norm": 2.0008212035639787,
"learning_rate": 1.4976616880546462e-05,
"loss": 0.5488,
"step": 70
},
{
"epoch": 0.18831837575400912,
"grad_norm": 2.1468051355877065,
"learning_rate": 1.4959113721231613e-05,
"loss": 0.5378,
"step": 80
},
{
"epoch": 0.21185817272326027,
"grad_norm": 2.09758107733118,
"learning_rate": 1.4936768422101134e-05,
"loss": 0.524,
"step": 90
},
{
"epoch": 0.2353979696925114,
"grad_norm": 1.9891368777701572,
"learning_rate": 1.4909595488774145e-05,
"loss": 0.5238,
"step": 100
},
{
"epoch": 0.25893776666176255,
"grad_norm": 1.8856649935655327,
"learning_rate": 1.487761256076484e-05,
"loss": 0.51,
"step": 110
},
{
"epoch": 0.2824775636310137,
"grad_norm": 1.9946450768120385,
"learning_rate": 1.4840840400031667e-05,
"loss": 0.5045,
"step": 120
},
{
"epoch": 0.3060173606002648,
"grad_norm": 2.02887277527658,
"learning_rate": 1.4799302877499543e-05,
"loss": 0.4993,
"step": 130
},
{
"epoch": 0.329557157569516,
"grad_norm": 1.7846741494593932,
"learning_rate": 1.475302695756387e-05,
"loss": 0.4959,
"step": 140
},
{
"epoch": 0.3530969545387671,
"grad_norm": 1.7910316946805966,
"learning_rate": 1.4702042680586378e-05,
"loss": 0.4903,
"step": 150
},
{
"epoch": 0.37663675150801823,
"grad_norm": 1.583559495202325,
"learning_rate": 1.4646383143394222e-05,
"loss": 0.4852,
"step": 160
},
{
"epoch": 0.40017654847726936,
"grad_norm": 1.8258988651876884,
"learning_rate": 1.458608447779491e-05,
"loss": 0.4835,
"step": 170
},
{
"epoch": 0.42371634544652054,
"grad_norm": 1.7923596937902033,
"learning_rate": 1.4521185827121071e-05,
"loss": 0.474,
"step": 180
},
{
"epoch": 0.44725614241577166,
"grad_norm": 1.8160198247756183,
"learning_rate": 1.4451729320820273e-05,
"loss": 0.4721,
"step": 190
},
{
"epoch": 0.4707959393850228,
"grad_norm": 1.7150789518558318,
"learning_rate": 1.437776004710637e-05,
"loss": 0.4729,
"step": 200
},
{
"epoch": 0.49433573635427397,
"grad_norm": 1.7400033525862848,
"learning_rate": 1.4299326023690146e-05,
"loss": 0.4607,
"step": 210
},
{
"epoch": 0.5178755333235251,
"grad_norm": 1.7002289663882648,
"learning_rate": 1.4216478166608262e-05,
"loss": 0.4613,
"step": 220
},
{
"epoch": 0.5414153302927762,
"grad_norm": 1.534972777417761,
"learning_rate": 1.4129270257170722e-05,
"loss": 0.4567,
"step": 230
},
{
"epoch": 0.5649551272620273,
"grad_norm": 1.6489960881148158,
"learning_rate": 1.403775890704834e-05,
"loss": 0.4559,
"step": 240
},
{
"epoch": 0.5884949242312785,
"grad_norm": 1.7846842786992185,
"learning_rate": 1.3942003521522837e-05,
"loss": 0.4624,
"step": 250
},
{
"epoch": 0.6120347212005296,
"grad_norm": 1.4351473554704604,
"learning_rate": 1.3842066260923462e-05,
"loss": 0.4496,
"step": 260
},
{
"epoch": 0.6355745181697808,
"grad_norm": 1.541683257911224,
"learning_rate": 1.3738012000275142e-05,
"loss": 0.4516,
"step": 270
},
{
"epoch": 0.659114315139032,
"grad_norm": 1.5879015034427175,
"learning_rate": 1.3629908287184381e-05,
"loss": 0.4475,
"step": 280
},
{
"epoch": 0.6826541121082831,
"grad_norm": 1.728309878285744,
"learning_rate": 1.3517825297990205e-05,
"loss": 0.4446,
"step": 290
},
{
"epoch": 0.7061939090775342,
"grad_norm": 1.4690146169046678,
"learning_rate": 1.340183579220868e-05,
"loss": 0.4491,
"step": 300
},
{
"epoch": 0.7297337060467853,
"grad_norm": 1.4969895208337947,
"learning_rate": 1.328201506530052e-05,
"loss": 0.4436,
"step": 310
},
{
"epoch": 0.7532735030160365,
"grad_norm": 1.4385449709842155,
"learning_rate": 1.3158440899792465e-05,
"loss": 0.4344,
"step": 320
},
{
"epoch": 0.7768132999852876,
"grad_norm": 1.4819229027772451,
"learning_rate": 1.3031193514784178e-05,
"loss": 0.4387,
"step": 330
},
{
"epoch": 0.8003530969545387,
"grad_norm": 1.5138328167674073,
"learning_rate": 1.2900355513873408e-05,
"loss": 0.4374,
"step": 340
},
{
"epoch": 0.82389289392379,
"grad_norm": 1.5486731806719658,
"learning_rate": 1.276601183153324e-05,
"loss": 0.4268,
"step": 350
},
{
"epoch": 0.8474326908930411,
"grad_norm": 1.3966287731811042,
"learning_rate": 1.2628249677976246e-05,
"loss": 0.4222,
"step": 360
},
{
"epoch": 0.8709724878622922,
"grad_norm": 1.4851117277296138,
"learning_rate": 1.2487158482541324e-05,
"loss": 0.4304,
"step": 370
},
{
"epoch": 0.8945122848315433,
"grad_norm": 1.4886042812511937,
"learning_rate": 1.2342829835639957e-05,
"loss": 0.4327,
"step": 380
},
{
"epoch": 0.9180520818007945,
"grad_norm": 1.5916975841301912,
"learning_rate": 1.219535742929962e-05,
"loss": 0.4312,
"step": 390
},
{
"epoch": 0.9415918787700456,
"grad_norm": 1.4359537454947529,
"learning_rate": 1.204483699634289e-05,
"loss": 0.418,
"step": 400
},
{
"epoch": 0.9651316757392967,
"grad_norm": 1.5039520247360716,
"learning_rate": 1.1891366248241758e-05,
"loss": 0.4169,
"step": 410
},
{
"epoch": 0.9886714727085479,
"grad_norm": 1.4396432279411446,
"learning_rate": 1.1735044811687508e-05,
"loss": 0.4165,
"step": 420
},
{
"epoch": 1.012211269677799,
"grad_norm": 1.4837321963480583,
"learning_rate": 1.1575974163917282e-05,
"loss": 0.3505,
"step": 430
},
{
"epoch": 1.0357510666470502,
"grad_norm": 1.438403278814582,
"learning_rate": 1.1414257566839376e-05,
"loss": 0.2795,
"step": 440
},
{
"epoch": 1.0592908636163012,
"grad_norm": 1.510877062877757,
"learning_rate": 1.125e-05,
"loss": 0.2814,
"step": 450
},
{
"epoch": 1.0828306605855524,
"grad_norm": 1.519349451859131,
"learning_rate": 1.1083308092435006e-05,
"loss": 0.2773,
"step": 460
},
{
"epoch": 1.1063704575548037,
"grad_norm": 1.3106222683180944,
"learning_rate": 1.091429005345085e-05,
"loss": 0.2706,
"step": 470
},
{
"epoch": 1.1299102545240547,
"grad_norm": 1.3984600885535579,
"learning_rate": 1.0743055602379712e-05,
"loss": 0.2754,
"step": 480
},
{
"epoch": 1.153450051493306,
"grad_norm": 1.4833593619381575,
"learning_rate": 1.0569715897354354e-05,
"loss": 0.2756,
"step": 490
},
{
"epoch": 1.176989848462557,
"grad_norm": 1.3932629119237105,
"learning_rate": 1.0394383463148993e-05,
"loss": 0.27,
"step": 500
},
{
"epoch": 1.2005296454318082,
"grad_norm": 1.4062493836827707,
"learning_rate": 1.0217172118132994e-05,
"loss": 0.2703,
"step": 510
},
{
"epoch": 1.2240694424010594,
"grad_norm": 1.4874549162201474,
"learning_rate": 1.003819690038481e-05,
"loss": 0.2739,
"step": 520
},
{
"epoch": 1.2476092393703104,
"grad_norm": 1.3708361425521698,
"learning_rate": 9.857573993014161e-06,
"loss": 0.2754,
"step": 530
},
{
"epoch": 1.2711490363395614,
"grad_norm": 1.3978327948674458,
"learning_rate": 9.675420648740886e-06,
"loss": 0.2692,
"step": 540
},
{
"epoch": 1.2946888333088127,
"grad_norm": 1.4715538000650579,
"learning_rate": 9.491855113779456e-06,
"loss": 0.2694,
"step": 550
},
{
"epoch": 1.318228630278064,
"grad_norm": 1.3468394374097452,
"learning_rate": 9.306996551078556e-06,
"loss": 0.2678,
"step": 560
},
{
"epoch": 1.341768427247315,
"grad_norm": 1.5175300555093134,
"learning_rate": 9.120964962965565e-06,
"loss": 0.266,
"step": 570
},
{
"epoch": 1.3653082242165662,
"grad_norm": 1.317724218624981,
"learning_rate": 8.933881113246134e-06,
"loss": 0.2666,
"step": 580
},
{
"epoch": 1.3888480211858174,
"grad_norm": 1.3809839276376867,
"learning_rate": 8.74586644880946e-06,
"loss": 0.2672,
"step": 590
},
{
"epoch": 1.4123878181550684,
"grad_norm": 1.3120307514577167,
"learning_rate": 8.557043020790113e-06,
"loss": 0.2671,
"step": 600
},
{
"epoch": 1.4359276151243194,
"grad_norm": 1.5865169519570839,
"learning_rate": 8.367533405337635e-06,
"loss": 0.2669,
"step": 610
},
{
"epoch": 1.4594674120935707,
"grad_norm": 1.315530128518467,
"learning_rate": 8.177460624045303e-06,
"loss": 0.2619,
"step": 620
},
{
"epoch": 1.483007209062822,
"grad_norm": 1.2729315130214236,
"learning_rate": 7.986948064089741e-06,
"loss": 0.2625,
"step": 630
},
{
"epoch": 1.506547006032073,
"grad_norm": 1.2890769997858393,
"learning_rate": 7.7961193981332e-06,
"loss": 0.2624,
"step": 640
},
{
"epoch": 1.5300868030013242,
"grad_norm": 1.344250246429933,
"learning_rate": 7.605098504040519e-06,
"loss": 0.2571,
"step": 650
},
{
"epoch": 1.5536265999705754,
"grad_norm": 1.3112449299655193,
"learning_rate": 7.414009384462882e-06,
"loss": 0.2603,
"step": 660
},
{
"epoch": 1.5771663969398264,
"grad_norm": 1.3722682469514345,
"learning_rate": 7.22297608634056e-06,
"loss": 0.2592,
"step": 670
},
{
"epoch": 1.6007061939090774,
"grad_norm": 1.3170860077023523,
"learning_rate": 7.032122620376899e-06,
"loss": 0.2535,
"step": 680
},
{
"epoch": 1.6242459908783287,
"grad_norm": 1.3106743856940115,
"learning_rate": 6.841572880535854e-06,
"loss": 0.2531,
"step": 690
},
{
"epoch": 1.64778578784758,
"grad_norm": 1.2798600011553958,
"learning_rate": 6.6514505636152814e-06,
"loss": 0.2548,
"step": 700
},
{
"epoch": 1.671325584816831,
"grad_norm": 1.3968866134114841,
"learning_rate": 6.461879088948227e-06,
"loss": 0.2507,
"step": 710
},
{
"epoch": 1.6948653817860821,
"grad_norm": 1.378986621794313,
"learning_rate": 6.272981518284342e-06,
"loss": 0.249,
"step": 720
},
{
"epoch": 1.7184051787553334,
"grad_norm": 1.2776388271568753,
"learning_rate": 6.084880475903424e-06,
"loss": 0.2451,
"step": 730
},
{
"epoch": 1.7419449757245844,
"grad_norm": 1.317908633074014,
"learning_rate": 5.897698069012938e-06,
"loss": 0.2484,
"step": 740
},
{
"epoch": 1.7654847726938354,
"grad_norm": 1.2702586608772435,
"learning_rate": 5.711555808481213e-06,
"loss": 0.2486,
"step": 750
},
{
"epoch": 1.7890245696630867,
"grad_norm": 1.3790950940287556,
"learning_rate": 5.526574529957729e-06,
"loss": 0.2451,
"step": 760
},
{
"epoch": 1.8125643666323379,
"grad_norm": 1.417794065597115,
"learning_rate": 5.342874315431765e-06,
"loss": 0.2442,
"step": 770
},
{
"epoch": 1.836104163601589,
"grad_norm": 1.3118057182766703,
"learning_rate": 5.160574415280257e-06,
"loss": 0.2438,
"step": 780
},
{
"epoch": 1.85964396057084,
"grad_norm": 1.380491802365371,
"learning_rate": 4.979793170855537e-06,
"loss": 0.2368,
"step": 790
},
{
"epoch": 1.8831837575400914,
"grad_norm": 1.3404717450970274,
"learning_rate": 4.800647937663126e-06,
"loss": 0.2393,
"step": 800
},
{
"epoch": 1.9067235545093424,
"grad_norm": 1.3084085289097462,
"learning_rate": 4.623255009179547e-06,
"loss": 0.2402,
"step": 810
},
{
"epoch": 1.9302633514785934,
"grad_norm": 1.2819422568439027,
"learning_rate": 4.4477295413595385e-06,
"loss": 0.2359,
"step": 820
},
{
"epoch": 1.9538031484478446,
"grad_norm": 1.2506810831080168,
"learning_rate": 4.274185477881681e-06,
"loss": 0.2365,
"step": 830
},
{
"epoch": 1.9773429454170959,
"grad_norm": 1.236732854007973,
"learning_rate": 4.102735476181019e-06,
"loss": 0.2308,
"step": 840
},
{
"epoch": 2.000882742386347,
"grad_norm": 1.2110215418076606,
"learning_rate": 3.933490834316633e-06,
"loss": 0.2284,
"step": 850
},
{
"epoch": 2.024422539355598,
"grad_norm": 1.4341116240165481,
"learning_rate": 3.7665614187216603e-06,
"loss": 0.128,
"step": 860
},
{
"epoch": 2.0479623363248494,
"grad_norm": 1.1448959546708688,
"learning_rate": 3.6020555928826935e-06,
"loss": 0.1201,
"step": 870
},
{
"epoch": 2.0715021332941004,
"grad_norm": 1.1891044799932367,
"learning_rate": 3.4400801469947734e-06,
"loss": 0.1196,
"step": 880
},
{
"epoch": 2.0950419302633514,
"grad_norm": 1.2226413834941532,
"learning_rate": 3.2807402286377675e-06,
"loss": 0.1183,
"step": 890
},
{
"epoch": 2.1185817272326024,
"grad_norm": 1.3069856648534286,
"learning_rate": 3.1241392745189985e-06,
"loss": 0.1181,
"step": 900
},
{
"epoch": 2.142121524201854,
"grad_norm": 1.1702804083458354,
"learning_rate": 2.970378943326543e-06,
"loss": 0.1166,
"step": 910
},
{
"epoch": 2.165661321171105,
"grad_norm": 1.2124612400300803,
"learning_rate": 2.8195590497367222e-06,
"loss": 0.1155,
"step": 920
},
{
"epoch": 2.189201118140356,
"grad_norm": 1.1611863697964633,
"learning_rate": 2.671777499618652e-06,
"loss": 0.1159,
"step": 930
},
{
"epoch": 2.2127409151096074,
"grad_norm": 1.1087717944945865,
"learning_rate": 2.527130226477926e-06,
"loss": 0.1168,
"step": 940
},
{
"epoch": 2.2362807120788584,
"grad_norm": 1.2482164891383611,
"learning_rate": 2.3857111291806348e-06,
"loss": 0.114,
"step": 950
},
{
"epoch": 2.2598205090481094,
"grad_norm": 1.1704503748862236,
"learning_rate": 2.2476120109982267e-06,
"loss": 0.1146,
"step": 960
},
{
"epoch": 2.283360306017361,
"grad_norm": 1.1831598264423941,
"learning_rate": 2.1129225200127196e-06,
"loss": 0.1142,
"step": 970
},
{
"epoch": 2.306900102986612,
"grad_norm": 1.1923513771478957,
"learning_rate": 1.981730090920969e-06,
"loss": 0.1121,
"step": 980
},
{
"epoch": 2.330439899955863,
"grad_norm": 1.1650306486853608,
"learning_rate": 1.8541198882757892e-06,
"loss": 0.1103,
"step": 990
},
{
"epoch": 2.353979696925114,
"grad_norm": 1.2474184378379198,
"learning_rate": 1.73017475120072e-06,
"loss": 0.1111,
"step": 1000
},
{
"epoch": 2.3775194938943653,
"grad_norm": 1.1410121185645685,
"learning_rate": 1.6099751396144142e-06,
"loss": 0.1101,
"step": 1010
},
{
"epoch": 2.4010592908636164,
"grad_norm": 1.1781449006163716,
"learning_rate": 1.493599081999453e-06,
"loss": 0.1085,
"step": 1020
},
{
"epoch": 2.4245990878328674,
"grad_norm": 1.095164172281657,
"learning_rate": 1.3811221247495784e-06,
"loss": 0.109,
"step": 1030
},
{
"epoch": 2.448138884802119,
"grad_norm": 1.1290274158456068,
"learning_rate": 1.2726172831281898e-06,
"loss": 0.1083,
"step": 1040
},
{
"epoch": 2.47167868177137,
"grad_norm": 1.1205826114027153,
"learning_rate": 1.1681549938699157e-06,
"loss": 0.1079,
"step": 1050
},
{
"epoch": 2.495218478740621,
"grad_norm": 1.237936841602485,
"learning_rate": 1.0678030694561044e-06,
"loss": 0.1084,
"step": 1060
},
{
"epoch": 2.518758275709872,
"grad_norm": 1.2748079747165908,
"learning_rate": 9.716266540938183e-07,
"loss": 0.1084,
"step": 1070
},
{
"epoch": 2.542298072679123,
"grad_norm": 1.1620787715820426,
"learning_rate": 8.796881814269941e-07,
"loss": 0.106,
"step": 1080
},
{
"epoch": 2.5658378696483743,
"grad_norm": 1.1814024146181927,
"learning_rate": 7.92047334007169e-07,
"loss": 0.1037,
"step": 1090
},
{
"epoch": 2.5893776666176254,
"grad_norm": 1.1411929584644598,
"learning_rate": 7.087610045501023e-07,
"loss": 0.1044,
"step": 1100
},
{
"epoch": 2.612917463586877,
"grad_norm": 1.1770255689497389,
"learning_rate": 6.29883259003445e-07,
"loss": 0.1051,
"step": 1110
},
{
"epoch": 2.636457260556128,
"grad_norm": 1.1667654713458508,
"learning_rate": 5.554653014494176e-07,
"loss": 0.1031,
"step": 1120
},
{
"epoch": 2.659997057525379,
"grad_norm": 1.238846596793772,
"learning_rate": 4.855554408652985e-07,
"loss": 0.1039,
"step": 1130
},
{
"epoch": 2.68353685449463,
"grad_norm": 1.0962748439341052,
"learning_rate": 4.201990597632832e-07,
"loss": 0.1032,
"step": 1140
},
{
"epoch": 2.707076651463881,
"grad_norm": 1.1228066231093363,
"learning_rate": 3.5943858473009237e-07,
"loss": 0.105,
"step": 1150
},
{
"epoch": 2.7306164484331323,
"grad_norm": 1.1681919060053207,
"learning_rate": 3.033134588854289e-07,
"loss": 0.1029,
"step": 1160
},
{
"epoch": 2.7541562454023834,
"grad_norm": 1.107535834090668,
"learning_rate": 2.518601162771755e-07,
"loss": 0.1048,
"step": 1170
},
{
"epoch": 2.777696042371635,
"grad_norm": 1.0985797806813495,
"learning_rate": 2.0511195822997013e-07,
"loss": 0.102,
"step": 1180
},
{
"epoch": 2.801235839340886,
"grad_norm": 1.1563052202999995,
"learning_rate": 1.6309933166247403e-07,
"loss": 0.1028,
"step": 1190
},
{
"epoch": 2.824775636310137,
"grad_norm": 1.1299722934234167,
"learning_rate": 1.258495093874454e-07,
"loss": 0.1006,
"step": 1200
},
{
"epoch": 2.848315433279388,
"grad_norm": 1.0731524173833442,
"learning_rate": 9.338667240738619e-08,
"loss": 0.1029,
"step": 1210
},
{
"epoch": 2.871855230248639,
"grad_norm": 1.0991693593720597,
"learning_rate": 6.573189421726466e-08,
"loss": 0.102,
"step": 1220
},
{
"epoch": 2.8953950272178903,
"grad_norm": 1.1501966264186627,
"learning_rate": 4.2903127124496454e-08,
"loss": 0.1032,
"step": 1230
},
{
"epoch": 2.9189348241871413,
"grad_norm": 1.1357187349706255,
"learning_rate": 2.4915190595068415e-08,
"loss": 0.1031,
"step": 1240
},
{
"epoch": 2.942474621156393,
"grad_norm": 1.1136651822369734,
"learning_rate": 1.1779761633370633e-08,
"loss": 0.0998,
"step": 1250
},
{
"epoch": 2.966014418125644,
"grad_norm": 1.1212278780119938,
"learning_rate": 3.50536720197997e-09,
"loss": 0.1005,
"step": 1260
},
{
"epoch": 2.989554215094895,
"grad_norm": 1.1503104789107468,
"learning_rate": 9.73786863237014e-11,
"loss": 0.1019,
"step": 1270
},
{
"epoch": 2.994262174488745,
"step": 1272,
"total_flos": 76511791349760.0,
"train_loss": 0.29241873632015297,
"train_runtime": 70414.7871,
"train_samples_per_second": 1.158,
"train_steps_per_second": 0.018
}
],
"logging_steps": 10,
"max_steps": 1272,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 76511791349760.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}