|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 6.0, |
|
"eval_steps": 500, |
|
"global_step": 2490, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.060350030175015085, |
|
"grad_norm": 0.17640693485736847, |
|
"learning_rate": 9.599999999999999e-05, |
|
"loss": 1.7813, |
|
"mean_token_accuracy": 0.6300852990150452, |
|
"num_tokens": 156656.0, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.12070006035003017, |
|
"grad_norm": 0.24193964898586273, |
|
"learning_rate": 0.00019599999999999997, |
|
"loss": 0.9197, |
|
"mean_token_accuracy": 0.768382026553154, |
|
"num_tokens": 282982.0, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.18105009052504525, |
|
"grad_norm": 0.16499435901641846, |
|
"learning_rate": 0.000296, |
|
"loss": 0.5906, |
|
"mean_token_accuracy": 0.8341364151239395, |
|
"num_tokens": 441181.0, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.24140012070006034, |
|
"grad_norm": 0.23563049733638763, |
|
"learning_rate": 0.0002999269005776963, |
|
"loss": 0.4832, |
|
"mean_token_accuracy": 0.8592967188358307, |
|
"num_tokens": 567644.0, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.30175015087507545, |
|
"grad_norm": 0.22556591033935547, |
|
"learning_rate": 0.0002996953705789175, |
|
"loss": 0.3612, |
|
"mean_token_accuracy": 0.8925437909364701, |
|
"num_tokens": 725987.0, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.3621001810500905, |
|
"grad_norm": 0.33429527282714844, |
|
"learning_rate": 0.00029930552794275785, |
|
"loss": 0.3126, |
|
"mean_token_accuracy": 0.9086851555109025, |
|
"num_tokens": 853185.0, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.4224502112251056, |
|
"grad_norm": 0.27340370416641235, |
|
"learning_rate": 0.0002987577849532824, |
|
"loss": 0.2343, |
|
"mean_token_accuracy": 0.9301495373249054, |
|
"num_tokens": 1011232.0, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.4828002414001207, |
|
"grad_norm": 0.2711191475391388, |
|
"learning_rate": 0.00029805272088449905, |
|
"loss": 0.2021, |
|
"mean_token_accuracy": 0.9406860828399658, |
|
"num_tokens": 1138074.0, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.5431502715751357, |
|
"grad_norm": 0.19240038096904755, |
|
"learning_rate": 0.00029719108138773827, |
|
"loss": 0.1508, |
|
"mean_token_accuracy": 0.9550948125123978, |
|
"num_tokens": 1293601.0, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.6035003017501509, |
|
"grad_norm": 0.27221739292144775, |
|
"learning_rate": 0.00029617377770307837, |
|
"loss": 0.1563, |
|
"mean_token_accuracy": 0.9542003554105759, |
|
"num_tokens": 1418074.0, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.663850331925166, |
|
"grad_norm": 0.25977134704589844, |
|
"learning_rate": 0.0002950018856956494, |
|
"loss": 0.1228, |
|
"mean_token_accuracy": 0.9640595990419388, |
|
"num_tokens": 1577856.0, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.724200362100181, |
|
"grad_norm": 0.2311161458492279, |
|
"learning_rate": 0.0002936766447178356, |
|
"loss": 0.1229, |
|
"mean_token_accuracy": 0.9646531140804291, |
|
"num_tokens": 1704393.0, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.7845503922751962, |
|
"grad_norm": 0.1375264674425125, |
|
"learning_rate": 0.0002921994562985788, |
|
"loss": 0.0972, |
|
"mean_token_accuracy": 0.9722524845600128, |
|
"num_tokens": 1860935.0, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.8449004224502112, |
|
"grad_norm": 0.2860631048679352, |
|
"learning_rate": 0.0002905718826611708, |
|
"loss": 0.0853, |
|
"mean_token_accuracy": 0.9756521546840667, |
|
"num_tokens": 1988266.0, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.9052504526252263, |
|
"grad_norm": 0.11123040318489075, |
|
"learning_rate": 0.00028879564507109946, |
|
"loss": 0.0814, |
|
"mean_token_accuracy": 0.9769123244285584, |
|
"num_tokens": 2146122.0, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.9656004828002414, |
|
"grad_norm": 0.239139586687088, |
|
"learning_rate": 0.0002868726220156981, |
|
"loss": 0.0696, |
|
"mean_token_accuracy": 0.9802538651227951, |
|
"num_tokens": 2273996.0, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.07211296260356903, |
|
"eval_mean_token_accuracy": 0.979879263285044, |
|
"eval_num_tokens": 2354180.0, |
|
"eval_runtime": 29.4819, |
|
"eval_samples_per_second": 12.516, |
|
"eval_steps_per_second": 6.275, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.024140012070006, |
|
"grad_norm": 0.13852086663246155, |
|
"learning_rate": 0.0002848048472175225, |
|
"loss": 0.0764, |
|
"mean_token_accuracy": 0.9782544571099822, |
|
"num_tokens": 2422077.0, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.0844900422450212, |
|
"grad_norm": 0.17753440141677856, |
|
"learning_rate": 0.00028259450748355637, |
|
"loss": 0.0527, |
|
"mean_token_accuracy": 0.9847306323051452, |
|
"num_tokens": 2564180.0, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.1448400724200363, |
|
"grad_norm": 0.13473442196846008, |
|
"learning_rate": 0.00028024394039252005, |
|
"loss": 0.0697, |
|
"mean_token_accuracy": 0.9803690612316132, |
|
"num_tokens": 2705737.0, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.2051901025950513, |
|
"grad_norm": 0.04998031258583069, |
|
"learning_rate": 0.0002777556318227281, |
|
"loss": 0.0452, |
|
"mean_token_accuracy": 0.987179564833641, |
|
"num_tokens": 2848782.0, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.2655401327700664, |
|
"grad_norm": 0.09412259608507156, |
|
"learning_rate": 0.00027513221332311073, |
|
"loss": 0.0615, |
|
"mean_token_accuracy": 0.9825534737110138, |
|
"num_tokens": 2991024.0, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.3258901629450814, |
|
"grad_norm": 0.12740211188793182, |
|
"learning_rate": 0.0002723764593301788, |
|
"loss": 0.0452, |
|
"mean_token_accuracy": 0.9870287185907364, |
|
"num_tokens": 3132622.0, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.3862401931200965, |
|
"grad_norm": 0.11777028441429138, |
|
"learning_rate": 0.0002694912842338756, |
|
"loss": 0.0568, |
|
"mean_token_accuracy": 0.9838440799713135, |
|
"num_tokens": 3273143.0, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.4465902232951118, |
|
"grad_norm": 0.09758122265338898, |
|
"learning_rate": 0.0002664797392954194, |
|
"loss": 0.0444, |
|
"mean_token_accuracy": 0.986908946633339, |
|
"num_tokens": 3416381.0, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.5069402534701268, |
|
"grad_norm": 0.0658789575099945, |
|
"learning_rate": 0.0002633450094203953, |
|
"loss": 0.0535, |
|
"mean_token_accuracy": 0.9848115313053131, |
|
"num_tokens": 3558249.0, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.567290283645142, |
|
"grad_norm": 0.07210762798786163, |
|
"learning_rate": 0.000260090409790509, |
|
"loss": 0.0409, |
|
"mean_token_accuracy": 0.9878959685564042, |
|
"num_tokens": 3700350.0, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.627640313820157, |
|
"grad_norm": 0.09006072580814362, |
|
"learning_rate": 0.000256719382357566, |
|
"loss": 0.052, |
|
"mean_token_accuracy": 0.9852559435367584, |
|
"num_tokens": 3842983.0, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.687990343995172, |
|
"grad_norm": 0.07121206820011139, |
|
"learning_rate": 0.0002532354922033823, |
|
"loss": 0.0401, |
|
"mean_token_accuracy": 0.9882262688875199, |
|
"num_tokens": 3985789.0, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.748340374170187, |
|
"grad_norm": 0.049150578677654266, |
|
"learning_rate": 0.00024964242376947747, |
|
"loss": 0.0514, |
|
"mean_token_accuracy": 0.9852595126628876, |
|
"num_tokens": 4128158.0, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.8086904043452021, |
|
"grad_norm": 0.08217272907495499, |
|
"learning_rate": 0.000245943976960537, |
|
"loss": 0.0373, |
|
"mean_token_accuracy": 0.9887230151891708, |
|
"num_tokens": 4270705.0, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.8690404345202172, |
|
"grad_norm": 0.060381677001714706, |
|
"learning_rate": 0.00024214406312576472, |
|
"loss": 0.051, |
|
"mean_token_accuracy": 0.9850554609298706, |
|
"num_tokens": 4412064.0, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.9293904646952322, |
|
"grad_norm": 0.07599000632762909, |
|
"learning_rate": 0.00023824670092237557, |
|
"loss": 0.0385, |
|
"mean_token_accuracy": 0.9883499753475189, |
|
"num_tokens": 4554646.0, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.9897404948702473, |
|
"grad_norm": 0.06878010928630829, |
|
"learning_rate": 0.00023425601206560257, |
|
"loss": 0.0432, |
|
"mean_token_accuracy": 0.9873087042570114, |
|
"num_tokens": 4688134.0, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.04542902857065201, |
|
"eval_mean_token_accuracy": 0.9870176924241556, |
|
"eval_num_tokens": 4708360.0, |
|
"eval_runtime": 29.4629, |
|
"eval_samples_per_second": 12.524, |
|
"eval_steps_per_second": 6.279, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.048280024140012, |
|
"grad_norm": 0.07973352819681168, |
|
"learning_rate": 0.00023017621696971407, |
|
"loss": 0.0424, |
|
"mean_token_accuracy": 0.9869058310371084, |
|
"num_tokens": 4837256.0, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.1086300543150274, |
|
"grad_norm": 0.3017246723175049, |
|
"learning_rate": 0.0002260116302846495, |
|
"loss": 0.0294, |
|
"mean_token_accuracy": 0.9908489334583283, |
|
"num_tokens": 4969729.0, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 2.1689800844900424, |
|
"grad_norm": 0.05929604917764664, |
|
"learning_rate": 0.0002217666563329952, |
|
"loss": 0.0407, |
|
"mean_token_accuracy": 0.9875844532251358, |
|
"num_tokens": 5120477.0, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.2293301146650575, |
|
"grad_norm": 0.10643558949232101, |
|
"learning_rate": 0.00021744578445212544, |
|
"loss": 0.03, |
|
"mean_token_accuracy": 0.9906252521276474, |
|
"num_tokens": 5253578.0, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 2.2896801448400725, |
|
"grad_norm": 0.07998061180114746, |
|
"learning_rate": 0.0002130535842464348, |
|
"loss": 0.0405, |
|
"mean_token_accuracy": 0.9873760217428207, |
|
"num_tokens": 5406645.0, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.3500301750150876, |
|
"grad_norm": 0.03785248100757599, |
|
"learning_rate": 0.0002085947007546829, |
|
"loss": 0.0286, |
|
"mean_token_accuracy": 0.9912718170881272, |
|
"num_tokens": 5540521.0, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 2.4103802051901027, |
|
"grad_norm": 0.038521163165569305, |
|
"learning_rate": 0.00020407384953756216, |
|
"loss": 0.0402, |
|
"mean_token_accuracy": 0.9876785135269165, |
|
"num_tokens": 5691357.0, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.4707302353651177, |
|
"grad_norm": 0.08188804239034653, |
|
"learning_rate": 0.00019949581169068456, |
|
"loss": 0.0286, |
|
"mean_token_accuracy": 0.991096887588501, |
|
"num_tokens": 5824399.0, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 2.5310802655401328, |
|
"grad_norm": 0.07109837234020233, |
|
"learning_rate": 0.0001948654287882601, |
|
"loss": 0.0388, |
|
"mean_token_accuracy": 0.9885295808315278, |
|
"num_tokens": 5975841.0, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.591430295715148, |
|
"grad_norm": 0.06275477260351181, |
|
"learning_rate": 0.00019018759776281605, |
|
"loss": 0.0261, |
|
"mean_token_accuracy": 0.9916540479660034, |
|
"num_tokens": 6108567.0, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 2.651780325890163, |
|
"grad_norm": 0.04038365185260773, |
|
"learning_rate": 0.00018546726572637065, |
|
"loss": 0.0352, |
|
"mean_token_accuracy": 0.9892910522222519, |
|
"num_tokens": 6259991.0, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.712130356065178, |
|
"grad_norm": 0.09150233864784241, |
|
"learning_rate": 0.00018070942473853873, |
|
"loss": 0.0255, |
|
"mean_token_accuracy": 0.9921514791250229, |
|
"num_tokens": 6393402.0, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 2.772480386240193, |
|
"grad_norm": 0.04401474818587303, |
|
"learning_rate": 0.00017591910652710262, |
|
"loss": 0.0355, |
|
"mean_token_accuracy": 0.9891881144046784, |
|
"num_tokens": 6544386.0, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 2.832830416415208, |
|
"grad_norm": 0.16860009729862213, |
|
"learning_rate": 0.00017110137716663107, |
|
"loss": 0.026, |
|
"mean_token_accuracy": 0.9918778198957443, |
|
"num_tokens": 6674740.0, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 2.8931804465902236, |
|
"grad_norm": 0.05165860429406166, |
|
"learning_rate": 0.0001662613317207742, |
|
"loss": 0.0403, |
|
"mean_token_accuracy": 0.9879327750205994, |
|
"num_tokens": 6826652.0, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 2.9535304767652386, |
|
"grad_norm": 0.06370134651660919, |
|
"learning_rate": 0.00016140408885390107, |
|
"loss": 0.0273, |
|
"mean_token_accuracy": 0.9915495270490646, |
|
"num_tokens": 6960481.0, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.038739174604415894, |
|
"eval_mean_token_accuracy": 0.9888446437345969, |
|
"eval_num_tokens": 7062540.0, |
|
"eval_runtime": 29.4618, |
|
"eval_samples_per_second": 12.525, |
|
"eval_steps_per_second": 6.279, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 3.012070006035003, |
|
"grad_norm": 0.05011030286550522, |
|
"learning_rate": 0.0001565347854177771, |
|
"loss": 0.0303, |
|
"mean_token_accuracy": 0.9901853019429236, |
|
"num_tokens": 7098331.0, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 3.0724200362100182, |
|
"grad_norm": 0.05125705525279045, |
|
"learning_rate": 0.00015165857101900816, |
|
"loss": 0.0233, |
|
"mean_token_accuracy": 0.9924298238754272, |
|
"num_tokens": 7246298.0, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 3.1327700663850333, |
|
"grad_norm": 0.059397757053375244, |
|
"learning_rate": 0.00014678060257299454, |
|
"loss": 0.027, |
|
"mean_token_accuracy": 0.991271983385086, |
|
"num_tokens": 7381191.0, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 3.1931200965600484, |
|
"grad_norm": 0.031168634071946144, |
|
"learning_rate": 0.00014190603885015624, |
|
"loss": 0.025, |
|
"mean_token_accuracy": 0.9918223685026168, |
|
"num_tokens": 7529646.0, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 3.2534701267350634, |
|
"grad_norm": 0.06622699648141861, |
|
"learning_rate": 0.00013704003502019595, |
|
"loss": 0.0274, |
|
"mean_token_accuracy": 0.9909968906641007, |
|
"num_tokens": 7665384.0, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 3.3138201569100785, |
|
"grad_norm": 0.040858324617147446, |
|
"learning_rate": 0.0001321877372001702, |
|
"loss": 0.0234, |
|
"mean_token_accuracy": 0.992581251859665, |
|
"num_tokens": 7814867.0, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 3.3741701870850935, |
|
"grad_norm": 0.06217949092388153, |
|
"learning_rate": 0.00012735427701213444, |
|
"loss": 0.0264, |
|
"mean_token_accuracy": 0.9915199714899063, |
|
"num_tokens": 7951694.0, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 3.4345202172601086, |
|
"grad_norm": 0.044205911457538605, |
|
"learning_rate": 0.00012254476615611694, |
|
"loss": 0.0229, |
|
"mean_token_accuracy": 0.9923870205879212, |
|
"num_tokens": 8099509.0, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 3.4948702474351236, |
|
"grad_norm": 0.042222440242767334, |
|
"learning_rate": 0.00011776429100416252, |
|
"loss": 0.0283, |
|
"mean_token_accuracy": 0.9907158309221268, |
|
"num_tokens": 8235602.0, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 3.5552202776101387, |
|
"grad_norm": 0.036372531205415726, |
|
"learning_rate": 0.00011301790722116113, |
|
"loss": 0.0227, |
|
"mean_token_accuracy": 0.9924897265434265, |
|
"num_tokens": 8384445.0, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 3.6155703077851538, |
|
"grad_norm": 0.06498222798109055, |
|
"learning_rate": 0.00010831063441815225, |
|
"loss": 0.0254, |
|
"mean_token_accuracy": 0.9921332412958145, |
|
"num_tokens": 8520372.0, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 3.675920337960169, |
|
"grad_norm": 0.03929189220070839, |
|
"learning_rate": 0.0001036474508437579, |
|
"loss": 0.0221, |
|
"mean_token_accuracy": 0.9927822852134705, |
|
"num_tokens": 8668318.0, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 3.736270368135184, |
|
"grad_norm": 0.06582839041948318, |
|
"learning_rate": 9.903328811935959e-05, |
|
"loss": 0.0245, |
|
"mean_token_accuracy": 0.9920618611574173, |
|
"num_tokens": 8805749.0, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 3.796620398310199, |
|
"grad_norm": 0.04050470143556595, |
|
"learning_rate": 9.447302602358619e-05, |
|
"loss": 0.0219, |
|
"mean_token_accuracy": 0.9927525413036347, |
|
"num_tokens": 8953240.0, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 3.856970428485214, |
|
"grad_norm": 0.06876744329929352, |
|
"learning_rate": 8.997148733162942e-05, |
|
"loss": 0.0263, |
|
"mean_token_accuracy": 0.991570799946785, |
|
"num_tokens": 9088072.0, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 3.9173204586602295, |
|
"grad_norm": 0.026705719530582428, |
|
"learning_rate": 8.553343271484368e-05, |
|
"loss": 0.0226, |
|
"mean_token_accuracy": 0.9928545600175858, |
|
"num_tokens": 9236720.0, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 3.9776704888352445, |
|
"grad_norm": 0.03888670355081558, |
|
"learning_rate": 8.116355570602482e-05, |
|
"loss": 0.0228, |
|
"mean_token_accuracy": 0.9928395706415176, |
|
"num_tokens": 9369354.0, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.03669499605894089, |
|
"eval_mean_token_accuracy": 0.9898246610486829, |
|
"eval_num_tokens": 9416720.0, |
|
"eval_runtime": 29.4478, |
|
"eval_samples_per_second": 12.531, |
|
"eval_steps_per_second": 6.282, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 4.036210018105009, |
|
"grad_norm": 0.028738977387547493, |
|
"learning_rate": 7.686647773569294e-05, |
|
"loss": 0.0238, |
|
"mean_token_accuracy": 0.9924964775744173, |
|
"num_tokens": 9515553.0, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 4.096560048280024, |
|
"grad_norm": 0.0359547957777977, |
|
"learning_rate": 7.264674324462724e-05, |
|
"loss": 0.0178, |
|
"mean_token_accuracy": 0.994020511507988, |
|
"num_tokens": 9653573.0, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 4.15691007845504, |
|
"grad_norm": 0.06054285541176796, |
|
"learning_rate": 6.850881487782298e-05, |
|
"loss": 0.0216, |
|
"mean_token_accuracy": 0.9927907025814057, |
|
"num_tokens": 9799592.0, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 4.217260108630055, |
|
"grad_norm": 0.02831297554075718, |
|
"learning_rate": 6.445706876495263e-05, |
|
"loss": 0.0176, |
|
"mean_token_accuracy": 0.9940158641338348, |
|
"num_tokens": 9937174.0, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 4.27761013880507, |
|
"grad_norm": 0.032845254987478256, |
|
"learning_rate": 6.0495789892323177e-05, |
|
"loss": 0.0209, |
|
"mean_token_accuracy": 0.9930106937885285, |
|
"num_tokens": 10084481.0, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 4.337960168980085, |
|
"grad_norm": 0.037183333188295364, |
|
"learning_rate": 5.6629167571222614e-05, |
|
"loss": 0.0176, |
|
"mean_token_accuracy": 0.994132205247879, |
|
"num_tokens": 10223690.0, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 4.3983101991551, |
|
"grad_norm": 0.04124921187758446, |
|
"learning_rate": 5.286129100744953e-05, |
|
"loss": 0.0231, |
|
"mean_token_accuracy": 0.992373656630516, |
|
"num_tokens": 10370189.0, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 4.458660229330115, |
|
"grad_norm": 0.046396173536777496, |
|
"learning_rate": 4.9196144976710996e-05, |
|
"loss": 0.0168, |
|
"mean_token_accuracy": 0.9944910758733749, |
|
"num_tokens": 10505942.0, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 4.51901025950513, |
|
"grad_norm": 0.0499492809176445, |
|
"learning_rate": 4.563760561046167e-05, |
|
"loss": 0.0205, |
|
"mean_token_accuracy": 0.9931335169076919, |
|
"num_tokens": 10652149.0, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 4.579360289680145, |
|
"grad_norm": 0.05738436430692673, |
|
"learning_rate": 4.2189436296641304e-05, |
|
"loss": 0.0173, |
|
"mean_token_accuracy": 0.9943641519546509, |
|
"num_tokens": 10789605.0, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 4.63971031985516, |
|
"grad_norm": 0.03117590956389904, |
|
"learning_rate": 3.885528369964654e-05, |
|
"loss": 0.0201, |
|
"mean_token_accuracy": 0.9931588870286941, |
|
"num_tokens": 10937347.0, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 4.700060350030175, |
|
"grad_norm": 0.03524640202522278, |
|
"learning_rate": 3.563867390374445e-05, |
|
"loss": 0.0179, |
|
"mean_token_accuracy": 0.9940504628419876, |
|
"num_tokens": 11076619.0, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 4.76041038020519, |
|
"grad_norm": 0.03088066540658474, |
|
"learning_rate": 3.254300868400823e-05, |
|
"loss": 0.0225, |
|
"mean_token_accuracy": 0.9924430441856384, |
|
"num_tokens": 11224409.0, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 4.820760410380205, |
|
"grad_norm": 0.04579736292362213, |
|
"learning_rate": 2.9571561908717783e-05, |
|
"loss": 0.0171, |
|
"mean_token_accuracy": 0.9944288891553879, |
|
"num_tokens": 11361190.0, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 4.88111044055522, |
|
"grad_norm": 0.043261025100946426, |
|
"learning_rate": 2.672747607703e-05, |
|
"loss": 0.0197, |
|
"mean_token_accuracy": 0.9936026883125305, |
|
"num_tokens": 11506394.0, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 4.941460470730235, |
|
"grad_norm": 0.04083774983882904, |
|
"learning_rate": 2.4013758995580522e-05, |
|
"loss": 0.0167, |
|
"mean_token_accuracy": 0.9945623755455018, |
|
"num_tokens": 11643469.0, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 0.11111137270927429, |
|
"learning_rate": 2.143328059753165e-05, |
|
"loss": 0.0179, |
|
"mean_token_accuracy": 0.994264479764958, |
|
"num_tokens": 11770900.0, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.037259791046381, |
|
"eval_mean_token_accuracy": 0.9904078212944237, |
|
"eval_num_tokens": 11770900.0, |
|
"eval_runtime": 29.4836, |
|
"eval_samples_per_second": 12.515, |
|
"eval_steps_per_second": 6.275, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 5.060350030175015, |
|
"grad_norm": 0.037272680550813675, |
|
"learning_rate": 1.8988769907430552e-05, |
|
"loss": 0.0157, |
|
"mean_token_accuracy": 0.9945630472898483, |
|
"num_tokens": 11928883.0, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 5.12070006035003, |
|
"grad_norm": 0.03758702054619789, |
|
"learning_rate": 1.6682812155087255e-05, |
|
"loss": 0.0154, |
|
"mean_token_accuracy": 0.9949316030740738, |
|
"num_tokens": 12056002.0, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 5.181050090525045, |
|
"grad_norm": 0.03220449388027191, |
|
"learning_rate": 1.4517846041525417e-05, |
|
"loss": 0.0167, |
|
"mean_token_accuracy": 0.9945937013626098, |
|
"num_tokens": 12216496.0, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 5.24140012070006, |
|
"grad_norm": 0.04776952043175697, |
|
"learning_rate": 1.2496161159896474e-05, |
|
"loss": 0.0162, |
|
"mean_token_accuracy": 0.9944632011651993, |
|
"num_tokens": 12343794.0, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 5.301750150875075, |
|
"grad_norm": 0.03377736359834671, |
|
"learning_rate": 1.061989557408528e-05, |
|
"loss": 0.017, |
|
"mean_token_accuracy": 0.9942396420240402, |
|
"num_tokens": 12500804.0, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 5.36210018105009, |
|
"grad_norm": 0.030239921063184738, |
|
"learning_rate": 8.891033557567768e-06, |
|
"loss": 0.0157, |
|
"mean_token_accuracy": 0.9947306287288665, |
|
"num_tokens": 12627390.0, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 5.422450211225105, |
|
"grad_norm": 0.03189610317349434, |
|
"learning_rate": 7.311403494912199e-06, |
|
"loss": 0.0178, |
|
"mean_token_accuracy": 0.9941246539354325, |
|
"num_tokens": 12785794.0, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 5.4828002414001205, |
|
"grad_norm": 0.04312798008322716, |
|
"learning_rate": 5.882675948142939e-06, |
|
"loss": 0.0156, |
|
"mean_token_accuracy": 0.9949129837751388, |
|
"num_tokens": 12911179.0, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 5.5431502715751355, |
|
"grad_norm": 0.03736381605267525, |
|
"learning_rate": 4.606361890011789e-06, |
|
"loss": 0.0182, |
|
"mean_token_accuracy": 0.9938953548669816, |
|
"num_tokens": 13067665.0, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 5.603500301750151, |
|
"grad_norm": 0.040421050041913986, |
|
"learning_rate": 3.4838111060457972e-06, |
|
"loss": 0.0157, |
|
"mean_token_accuracy": 0.9949649393558502, |
|
"num_tokens": 13195440.0, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 5.663850331925166, |
|
"grad_norm": 0.0452859103679657, |
|
"learning_rate": 2.5162107670607335e-06, |
|
"loss": 0.0157, |
|
"mean_token_accuracy": 0.9947731006145477, |
|
"num_tokens": 13351533.0, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 5.724200362100181, |
|
"grad_norm": 0.032347723841667175, |
|
"learning_rate": 1.7045841736502918e-06, |
|
"loss": 0.0159, |
|
"mean_token_accuracy": 0.9948224556446076, |
|
"num_tokens": 13477597.0, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 5.784550392275197, |
|
"grad_norm": 0.029396912083029747, |
|
"learning_rate": 1.0497896739790346e-06, |
|
"loss": 0.0161, |
|
"mean_token_accuracy": 0.9947206151485443, |
|
"num_tokens": 13634305.0, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 5.844900422450211, |
|
"grad_norm": 0.03758098930120468, |
|
"learning_rate": 5.525197560229233e-07, |
|
"loss": 0.0162, |
|
"mean_token_accuracy": 0.9944813293218613, |
|
"num_tokens": 13760787.0, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 5.905250452625227, |
|
"grad_norm": 0.035603396594524384, |
|
"learning_rate": 2.1330031521810676e-07, |
|
"loss": 0.0158, |
|
"mean_token_accuracy": 0.9946977686882019, |
|
"num_tokens": 13917086.0, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 5.965600482800241, |
|
"grad_norm": 0.0416802354156971, |
|
"learning_rate": 3.249009829212612e-08, |
|
"loss": 0.0159, |
|
"mean_token_accuracy": 0.9947532343864441, |
|
"num_tokens": 14043219.0, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.03885212540626526, |
|
"eval_mean_token_accuracy": 0.9902579658740276, |
|
"eval_num_tokens": 14125080.0, |
|
"eval_runtime": 29.4371, |
|
"eval_samples_per_second": 12.535, |
|
"eval_steps_per_second": 6.285, |
|
"step": 2490 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 2490, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 6, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.1988961755335332e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|