|
{ |
|
"best_metric": 0.3944380581378937, |
|
"best_model_checkpoint": "./FT_models/[LDH]0226_origin_data/checkpoint-3000", |
|
"epoch": 2.8483985765124555, |
|
"eval_steps": 500, |
|
"global_step": 3000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.009489916963226572, |
|
"grad_norm": 0.9249091148376465, |
|
"learning_rate": 0.0001999995055317446, |
|
"loss": 2.3789, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.018979833926453145, |
|
"grad_norm": 0.7498548030853271, |
|
"learning_rate": 0.0001999955498150411, |
|
"loss": 1.3794, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.028469750889679714, |
|
"grad_norm": 0.7051133513450623, |
|
"learning_rate": 0.00019998763853811184, |
|
"loss": 1.2474, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03795966785290629, |
|
"grad_norm": 0.7239620685577393, |
|
"learning_rate": 0.00019997577201390606, |
|
"loss": 1.2512, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04744958481613286, |
|
"grad_norm": 0.5843775272369385, |
|
"learning_rate": 0.0001999599507118322, |
|
"loss": 1.1068, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05693950177935943, |
|
"grad_norm": 0.59147709608078, |
|
"learning_rate": 0.00019994017525773913, |
|
"loss": 1.1164, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.066429418742586, |
|
"grad_norm": 0.6587395668029785, |
|
"learning_rate": 0.0001999164464338918, |
|
"loss": 1.063, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.07591933570581258, |
|
"grad_norm": 0.5964196920394897, |
|
"learning_rate": 0.0001998887651789398, |
|
"loss": 1.0046, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.08540925266903915, |
|
"grad_norm": 0.6036216020584106, |
|
"learning_rate": 0.0001998571325878806, |
|
"loss": 1.0529, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09489916963226572, |
|
"grad_norm": 0.7564125061035156, |
|
"learning_rate": 0.00019982154991201608, |
|
"loss": 1.0009, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.10438908659549229, |
|
"grad_norm": 0.7184653282165527, |
|
"learning_rate": 0.00019978201855890308, |
|
"loss": 0.9723, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.11387900355871886, |
|
"grad_norm": 0.6793704628944397, |
|
"learning_rate": 0.00019973854009229763, |
|
"loss": 0.9341, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.12336892052194544, |
|
"grad_norm": 0.6832011342048645, |
|
"learning_rate": 0.00019969111623209323, |
|
"loss": 0.9873, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.132858837485172, |
|
"grad_norm": 0.85871422290802, |
|
"learning_rate": 0.00019963974885425266, |
|
"loss": 0.9299, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.1423487544483986, |
|
"grad_norm": 0.7056658267974854, |
|
"learning_rate": 0.00019958443999073397, |
|
"loss": 0.9533, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.15183867141162516, |
|
"grad_norm": 0.7040665149688721, |
|
"learning_rate": 0.00019952519182940993, |
|
"loss": 0.8815, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.16132858837485173, |
|
"grad_norm": 0.6971840262413025, |
|
"learning_rate": 0.0001994620067139815, |
|
"loss": 0.9461, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.1708185053380783, |
|
"grad_norm": 0.7504563331604004, |
|
"learning_rate": 0.00019939488714388524, |
|
"loss": 0.8483, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.18030842230130487, |
|
"grad_norm": 0.684840977191925, |
|
"learning_rate": 0.00019932383577419432, |
|
"loss": 0.9122, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.18979833926453143, |
|
"grad_norm": 0.6745721101760864, |
|
"learning_rate": 0.0001992488554155135, |
|
"loss": 0.8271, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.199288256227758, |
|
"grad_norm": 0.7717174887657166, |
|
"learning_rate": 0.0001991699490338681, |
|
"loss": 0.8124, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.20877817319098457, |
|
"grad_norm": 0.7041401267051697, |
|
"learning_rate": 0.00019908711975058637, |
|
"loss": 0.8219, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.21826809015421114, |
|
"grad_norm": 0.6780515909194946, |
|
"learning_rate": 0.00019900037084217637, |
|
"loss": 0.7733, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.2277580071174377, |
|
"grad_norm": 0.7086532115936279, |
|
"learning_rate": 0.00019890970574019617, |
|
"loss": 0.7575, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.2372479240806643, |
|
"grad_norm": 0.7180762887001038, |
|
"learning_rate": 0.00019881512803111796, |
|
"loss": 0.8863, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.24673784104389088, |
|
"grad_norm": 0.8088333010673523, |
|
"learning_rate": 0.00019871664145618657, |
|
"loss": 0.7497, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.25622775800711745, |
|
"grad_norm": 0.6337082386016846, |
|
"learning_rate": 0.00019861424991127115, |
|
"loss": 0.8269, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.265717674970344, |
|
"grad_norm": 0.8026911020278931, |
|
"learning_rate": 0.00019850795744671116, |
|
"loss": 0.8229, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.2752075919335706, |
|
"grad_norm": 0.8139302730560303, |
|
"learning_rate": 0.00019839776826715614, |
|
"loss": 0.78, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.2846975088967972, |
|
"grad_norm": 0.8401020765304565, |
|
"learning_rate": 0.00019828368673139947, |
|
"loss": 0.6933, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.2941874258600237, |
|
"grad_norm": 0.7387434840202332, |
|
"learning_rate": 0.00019816571735220583, |
|
"loss": 0.8241, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.3036773428232503, |
|
"grad_norm": 0.7867814302444458, |
|
"learning_rate": 0.0001980438647961327, |
|
"loss": 0.695, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.31316725978647686, |
|
"grad_norm": 0.6026145219802856, |
|
"learning_rate": 0.00019791813388334581, |
|
"loss": 0.8118, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.32265717674970346, |
|
"grad_norm": 0.7333958148956299, |
|
"learning_rate": 0.00019778852958742853, |
|
"loss": 0.7278, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.33214709371293, |
|
"grad_norm": 0.7646610140800476, |
|
"learning_rate": 0.00019765505703518496, |
|
"loss": 0.6802, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.3416370106761566, |
|
"grad_norm": 0.8131152987480164, |
|
"learning_rate": 0.00019751772150643722, |
|
"loss": 0.766, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.35112692763938314, |
|
"grad_norm": 0.7349848747253418, |
|
"learning_rate": 0.0001973765284338167, |
|
"loss": 0.6832, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.36061684460260973, |
|
"grad_norm": 0.6821774244308472, |
|
"learning_rate": 0.00019723148340254892, |
|
"loss": 0.7976, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.3701067615658363, |
|
"grad_norm": 0.6879362463951111, |
|
"learning_rate": 0.0001970825921502328, |
|
"loss": 0.7195, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.37959667852906287, |
|
"grad_norm": 0.699598491191864, |
|
"learning_rate": 0.00019692986056661356, |
|
"loss": 0.7, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.38908659549228947, |
|
"grad_norm": 0.6818081736564636, |
|
"learning_rate": 0.0001967732946933499, |
|
"loss": 0.6991, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.398576512455516, |
|
"grad_norm": 0.6942121386528015, |
|
"learning_rate": 0.00019661290072377482, |
|
"loss": 0.7161, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.4080664294187426, |
|
"grad_norm": 0.6413360834121704, |
|
"learning_rate": 0.0001964486850026507, |
|
"loss": 0.7084, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.41755634638196915, |
|
"grad_norm": 0.6501904129981995, |
|
"learning_rate": 0.00019628065402591845, |
|
"loss": 0.7431, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.42704626334519574, |
|
"grad_norm": 0.6739936470985413, |
|
"learning_rate": 0.0001961088144404403, |
|
"loss": 0.7081, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.4365361803084223, |
|
"grad_norm": 0.7370271682739258, |
|
"learning_rate": 0.00019593317304373705, |
|
"loss": 0.7203, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.4460260972716489, |
|
"grad_norm": 0.5882676243782043, |
|
"learning_rate": 0.00019575373678371909, |
|
"loss": 0.679, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.4555160142348754, |
|
"grad_norm": 0.6047580242156982, |
|
"learning_rate": 0.0001955705127584117, |
|
"loss": 0.6712, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.465005931198102, |
|
"grad_norm": 0.6153593063354492, |
|
"learning_rate": 0.00019538350821567404, |
|
"loss": 0.6788, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.4744958481613286, |
|
"grad_norm": 0.6934164762496948, |
|
"learning_rate": 0.00019519273055291266, |
|
"loss": 0.6903, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.4744958481613286, |
|
"eval_loss": 0.6663702130317688, |
|
"eval_runtime": 20.5638, |
|
"eval_samples_per_second": 15.415, |
|
"eval_steps_per_second": 7.732, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.48398576512455516, |
|
"grad_norm": 0.6835016012191772, |
|
"learning_rate": 0.00019499818731678873, |
|
"loss": 0.6843, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.49347568208778175, |
|
"grad_norm": 0.666349470615387, |
|
"learning_rate": 0.00019479988620291956, |
|
"loss": 0.6961, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.5029655990510083, |
|
"grad_norm": 0.7474802732467651, |
|
"learning_rate": 0.00019459783505557424, |
|
"loss": 0.6614, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.5124555160142349, |
|
"grad_norm": 0.6815286874771118, |
|
"learning_rate": 0.0001943920418673633, |
|
"loss": 0.6693, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.5219454329774614, |
|
"grad_norm": 0.5686200261116028, |
|
"learning_rate": 0.0001941825147789225, |
|
"loss": 0.6946, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.531435349940688, |
|
"grad_norm": 0.643159806728363, |
|
"learning_rate": 0.00019396926207859084, |
|
"loss": 0.6486, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.5409252669039146, |
|
"grad_norm": 0.7110168933868408, |
|
"learning_rate": 0.00019375229220208276, |
|
"loss": 0.6459, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.5504151838671412, |
|
"grad_norm": 0.6624748706817627, |
|
"learning_rate": 0.0001935316137321543, |
|
"loss": 0.6692, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.5599051008303677, |
|
"grad_norm": 0.761031985282898, |
|
"learning_rate": 0.00019330723539826375, |
|
"loss": 0.6046, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.5693950177935944, |
|
"grad_norm": 0.6096632480621338, |
|
"learning_rate": 0.0001930791660762262, |
|
"loss": 0.6775, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5788849347568209, |
|
"grad_norm": 0.6898393630981445, |
|
"learning_rate": 0.0001928474147878626, |
|
"loss": 0.6293, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.5883748517200474, |
|
"grad_norm": 0.6664237380027771, |
|
"learning_rate": 0.0001926119907006426, |
|
"loss": 0.6274, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.597864768683274, |
|
"grad_norm": 0.7034701704978943, |
|
"learning_rate": 0.00019237290312732226, |
|
"loss": 0.6334, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.6073546856465006, |
|
"grad_norm": 0.5845565795898438, |
|
"learning_rate": 0.0001921301615255754, |
|
"loss": 0.6791, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.6168446026097272, |
|
"grad_norm": 0.6939854025840759, |
|
"learning_rate": 0.00019188377549761963, |
|
"loss": 0.6554, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.6263345195729537, |
|
"grad_norm": 0.6801786422729492, |
|
"learning_rate": 0.00019163375478983632, |
|
"loss": 0.6543, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.6358244365361803, |
|
"grad_norm": 0.583102285861969, |
|
"learning_rate": 0.00019138010929238534, |
|
"loss": 0.6268, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.6453143534994069, |
|
"grad_norm": 0.6107549667358398, |
|
"learning_rate": 0.0001911228490388136, |
|
"loss": 0.6274, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.6548042704626335, |
|
"grad_norm": 0.6578108072280884, |
|
"learning_rate": 0.00019086198420565823, |
|
"loss": 0.6171, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.66429418742586, |
|
"grad_norm": 0.6483505368232727, |
|
"learning_rate": 0.000190597525112044, |
|
"loss": 0.5673, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6737841043890866, |
|
"grad_norm": 0.698631763458252, |
|
"learning_rate": 0.00019032948221927524, |
|
"loss": 0.5675, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.6832740213523132, |
|
"grad_norm": 0.6782714128494263, |
|
"learning_rate": 0.00019005786613042185, |
|
"loss": 0.5904, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.6927639383155397, |
|
"grad_norm": 0.673217236995697, |
|
"learning_rate": 0.00018978268758989991, |
|
"loss": 0.622, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.7022538552787663, |
|
"grad_norm": 0.6975003480911255, |
|
"learning_rate": 0.00018950395748304678, |
|
"loss": 0.6558, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.7117437722419929, |
|
"grad_norm": 0.6466664671897888, |
|
"learning_rate": 0.0001892216868356904, |
|
"loss": 0.673, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.7212336892052195, |
|
"grad_norm": 0.6829777359962463, |
|
"learning_rate": 0.00018893588681371303, |
|
"loss": 0.5808, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.730723606168446, |
|
"grad_norm": 0.6941152811050415, |
|
"learning_rate": 0.00018864656872260985, |
|
"loss": 0.6232, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.7402135231316725, |
|
"grad_norm": 0.7243316769599915, |
|
"learning_rate": 0.00018835374400704154, |
|
"loss": 0.5933, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.7497034400948992, |
|
"grad_norm": 0.6251444816589355, |
|
"learning_rate": 0.00018805742425038145, |
|
"loss": 0.6116, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.7591933570581257, |
|
"grad_norm": 0.646567165851593, |
|
"learning_rate": 0.00018775762117425777, |
|
"loss": 0.601, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7686832740213523, |
|
"grad_norm": 0.6085071563720703, |
|
"learning_rate": 0.00018745434663808942, |
|
"loss": 0.6603, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.7781731909845789, |
|
"grad_norm": 0.6589152812957764, |
|
"learning_rate": 0.00018714761263861728, |
|
"loss": 0.5672, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.7876631079478055, |
|
"grad_norm": 0.7008639574050903, |
|
"learning_rate": 0.00018683743130942928, |
|
"loss": 0.6363, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.797153024911032, |
|
"grad_norm": 0.6070975065231323, |
|
"learning_rate": 0.00018652381492048083, |
|
"loss": 0.6075, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.8066429418742586, |
|
"grad_norm": 0.7204874753952026, |
|
"learning_rate": 0.00018620677587760916, |
|
"loss": 0.5797, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.8161328588374852, |
|
"grad_norm": 0.6779175400733948, |
|
"learning_rate": 0.00018588632672204264, |
|
"loss": 0.5729, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.8256227758007118, |
|
"grad_norm": 0.5760260224342346, |
|
"learning_rate": 0.00018556248012990468, |
|
"loss": 0.5568, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.8351126927639383, |
|
"grad_norm": 0.6586163640022278, |
|
"learning_rate": 0.0001852352489117124, |
|
"loss": 0.5734, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.8446026097271648, |
|
"grad_norm": 0.6545649766921997, |
|
"learning_rate": 0.0001849046460118698, |
|
"loss": 0.6065, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.8540925266903915, |
|
"grad_norm": 0.5810338258743286, |
|
"learning_rate": 0.00018457068450815562, |
|
"loss": 0.5435, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.863582443653618, |
|
"grad_norm": 0.5936874151229858, |
|
"learning_rate": 0.00018423337761120618, |
|
"loss": 0.5878, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.8730723606168446, |
|
"grad_norm": 0.6365971565246582, |
|
"learning_rate": 0.00018389273866399275, |
|
"loss": 0.5696, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.8825622775800712, |
|
"grad_norm": 0.5335951447486877, |
|
"learning_rate": 0.00018354878114129367, |
|
"loss": 0.5848, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.8920521945432978, |
|
"grad_norm": 0.6080604791641235, |
|
"learning_rate": 0.00018320151864916135, |
|
"loss": 0.5739, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.9015421115065243, |
|
"grad_norm": 0.5476921200752258, |
|
"learning_rate": 0.00018285096492438424, |
|
"loss": 0.5474, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.9110320284697508, |
|
"grad_norm": 0.6528366804122925, |
|
"learning_rate": 0.00018249713383394303, |
|
"loss": 0.542, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.9205219454329775, |
|
"grad_norm": 0.5690432786941528, |
|
"learning_rate": 0.00018214003937446253, |
|
"loss": 0.5771, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.930011862396204, |
|
"grad_norm": 0.7033817172050476, |
|
"learning_rate": 0.0001817796956716578, |
|
"loss": 0.537, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.9395017793594306, |
|
"grad_norm": 0.6648123264312744, |
|
"learning_rate": 0.00018141611697977529, |
|
"loss": 0.5579, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.9489916963226572, |
|
"grad_norm": 0.5434956550598145, |
|
"learning_rate": 0.0001810493176810292, |
|
"loss": 0.58, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.9489916963226572, |
|
"eval_loss": 0.5363849401473999, |
|
"eval_runtime": 20.4233, |
|
"eval_samples_per_second": 15.521, |
|
"eval_steps_per_second": 7.785, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.9584816132858838, |
|
"grad_norm": 0.6258216500282288, |
|
"learning_rate": 0.00018067931228503246, |
|
"loss": 0.6065, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.9679715302491103, |
|
"grad_norm": 0.49021461606025696, |
|
"learning_rate": 0.00018030611542822257, |
|
"loss": 0.5184, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.9774614472123369, |
|
"grad_norm": 0.6789380311965942, |
|
"learning_rate": 0.00017992974187328305, |
|
"loss": 0.558, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.9869513641755635, |
|
"grad_norm": 0.6128289103507996, |
|
"learning_rate": 0.000179550206508559, |
|
"loss": 0.5901, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.99644128113879, |
|
"grad_norm": 0.5243271589279175, |
|
"learning_rate": 0.00017916752434746856, |
|
"loss": 0.5455, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.0066429418742586, |
|
"grad_norm": 0.6114349365234375, |
|
"learning_rate": 0.00017878171052790868, |
|
"loss": 0.5706, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.0161328588374852, |
|
"grad_norm": 0.5785910487174988, |
|
"learning_rate": 0.00017839278031165658, |
|
"loss": 0.4914, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.0256227758007117, |
|
"grad_norm": 0.6278082132339478, |
|
"learning_rate": 0.00017800074908376584, |
|
"loss": 0.4585, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.0351126927639382, |
|
"grad_norm": 0.5266045331954956, |
|
"learning_rate": 0.0001776056323519579, |
|
"loss": 0.4257, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.0446026097271648, |
|
"grad_norm": 0.5933576226234436, |
|
"learning_rate": 0.00017720744574600863, |
|
"loss": 0.4579, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.0540925266903916, |
|
"grad_norm": 0.568048357963562, |
|
"learning_rate": 0.00017680620501712996, |
|
"loss": 0.4736, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.063582443653618, |
|
"grad_norm": 0.586031436920166, |
|
"learning_rate": 0.00017640192603734692, |
|
"loss": 0.4208, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.0730723606168446, |
|
"grad_norm": 0.632996141910553, |
|
"learning_rate": 0.00017599462479886974, |
|
"loss": 0.4518, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.0825622775800712, |
|
"grad_norm": 0.6920955777168274, |
|
"learning_rate": 0.00017558431741346122, |
|
"loss": 0.4593, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.0920521945432977, |
|
"grad_norm": 0.5045620799064636, |
|
"learning_rate": 0.00017517102011179933, |
|
"loss": 0.472, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.1015421115065243, |
|
"grad_norm": 0.5956369638442993, |
|
"learning_rate": 0.00017475474924283536, |
|
"loss": 0.446, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.1110320284697508, |
|
"grad_norm": 0.7154461741447449, |
|
"learning_rate": 0.000174335521273147, |
|
"loss": 0.4513, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.1205219454329776, |
|
"grad_norm": 0.5512189269065857, |
|
"learning_rate": 0.00017391335278628712, |
|
"loss": 0.4584, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.130011862396204, |
|
"grad_norm": 0.9251359105110168, |
|
"learning_rate": 0.0001734882604821276, |
|
"loss": 0.4276, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.1395017793594306, |
|
"grad_norm": 0.640958845615387, |
|
"learning_rate": 0.00017306026117619889, |
|
"loss": 0.4481, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.1489916963226572, |
|
"grad_norm": 0.5504414439201355, |
|
"learning_rate": 0.00017262937179902472, |
|
"loss": 0.4697, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.1584816132858837, |
|
"grad_norm": 0.575148344039917, |
|
"learning_rate": 0.00017219560939545246, |
|
"loss": 0.4614, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.1679715302491103, |
|
"grad_norm": 0.5111013650894165, |
|
"learning_rate": 0.0001717589911239788, |
|
"loss": 0.4314, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.1774614472123368, |
|
"grad_norm": 0.5844498872756958, |
|
"learning_rate": 0.00017131953425607104, |
|
"loss": 0.4628, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.1869513641755636, |
|
"grad_norm": 0.5282189846038818, |
|
"learning_rate": 0.00017087725617548385, |
|
"loss": 0.443, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.1964412811387901, |
|
"grad_norm": 0.5334432721138, |
|
"learning_rate": 0.00017043217437757164, |
|
"loss": 0.4927, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.2059311981020167, |
|
"grad_norm": 0.8024120926856995, |
|
"learning_rate": 0.00016998430646859654, |
|
"loss": 0.4808, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.2154211150652432, |
|
"grad_norm": 0.5527599453926086, |
|
"learning_rate": 0.00016953367016503182, |
|
"loss": 0.433, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.2249110320284697, |
|
"grad_norm": 0.6859824657440186, |
|
"learning_rate": 0.00016908028329286112, |
|
"loss": 0.4774, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.2344009489916963, |
|
"grad_norm": 0.592665433883667, |
|
"learning_rate": 0.0001686241637868734, |
|
"loss": 0.4595, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.2438908659549228, |
|
"grad_norm": 0.643675684928894, |
|
"learning_rate": 0.00016816532968995328, |
|
"loss": 0.4377, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.2533807829181494, |
|
"grad_norm": 0.5149793028831482, |
|
"learning_rate": 0.00016770379915236766, |
|
"loss": 0.4127, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.262870699881376, |
|
"grad_norm": 0.6323621273040771, |
|
"learning_rate": 0.00016723959043104728, |
|
"loss": 0.4821, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.2723606168446027, |
|
"grad_norm": 0.6195471882820129, |
|
"learning_rate": 0.00016677272188886483, |
|
"loss": 0.479, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.2818505338078292, |
|
"grad_norm": 0.6360680460929871, |
|
"learning_rate": 0.00016630321199390867, |
|
"loss": 0.4322, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.2913404507710557, |
|
"grad_norm": 0.614509642124176, |
|
"learning_rate": 0.00016583107931875192, |
|
"loss": 0.4992, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.3008303677342823, |
|
"grad_norm": 0.6960734724998474, |
|
"learning_rate": 0.00016535634253971794, |
|
"loss": 0.4765, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.3103202846975088, |
|
"grad_norm": 0.5039885640144348, |
|
"learning_rate": 0.00016487902043614173, |
|
"loss": 0.4437, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.3198102016607356, |
|
"grad_norm": 0.629076361656189, |
|
"learning_rate": 0.00016439913188962685, |
|
"loss": 0.4626, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.3293001186239621, |
|
"grad_norm": 0.538341760635376, |
|
"learning_rate": 0.0001639166958832985, |
|
"loss": 0.4166, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.3387900355871887, |
|
"grad_norm": 0.6629465222358704, |
|
"learning_rate": 0.00016343173150105278, |
|
"loss": 0.4736, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.3482799525504152, |
|
"grad_norm": 0.6603752374649048, |
|
"learning_rate": 0.0001629442579268016, |
|
"loss": 0.4522, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.3577698695136418, |
|
"grad_norm": 0.7585862874984741, |
|
"learning_rate": 0.0001624542944437139, |
|
"loss": 0.439, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.3672597864768683, |
|
"grad_norm": 0.5442929863929749, |
|
"learning_rate": 0.00016196186043345288, |
|
"loss": 0.4181, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.3767497034400948, |
|
"grad_norm": 0.5769705772399902, |
|
"learning_rate": 0.00016146697537540924, |
|
"loss": 0.4431, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.3862396204033214, |
|
"grad_norm": 0.6381351351737976, |
|
"learning_rate": 0.0001609696588459307, |
|
"loss": 0.4476, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.395729537366548, |
|
"grad_norm": 0.573300838470459, |
|
"learning_rate": 0.00016046993051754756, |
|
"loss": 0.4839, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.4052194543297747, |
|
"grad_norm": 0.607252836227417, |
|
"learning_rate": 0.0001599678101581945, |
|
"loss": 0.4528, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.4147093712930012, |
|
"grad_norm": 0.6059923768043518, |
|
"learning_rate": 0.00015946331763042867, |
|
"loss": 0.4493, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.4241992882562278, |
|
"grad_norm": 0.6414983868598938, |
|
"learning_rate": 0.00015895647289064396, |
|
"loss": 0.4455, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.4241992882562278, |
|
"eval_loss": 0.4805718660354614, |
|
"eval_runtime": 20.4306, |
|
"eval_samples_per_second": 15.516, |
|
"eval_steps_per_second": 7.782, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.4336892052194543, |
|
"grad_norm": 0.6877856850624084, |
|
"learning_rate": 0.0001584472959882815, |
|
"loss": 0.4553, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.4431791221826809, |
|
"grad_norm": 0.6739844679832458, |
|
"learning_rate": 0.0001579358070650367, |
|
"loss": 0.4682, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.4526690391459074, |
|
"grad_norm": 0.6030142903327942, |
|
"learning_rate": 0.00015742202635406235, |
|
"loss": 0.4521, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.4621589561091342, |
|
"grad_norm": 0.6368103623390198, |
|
"learning_rate": 0.0001569059741791684, |
|
"loss": 0.4458, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.4716488730723607, |
|
"grad_norm": 0.649683952331543, |
|
"learning_rate": 0.0001563876709540178, |
|
"loss": 0.4568, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.4811387900355872, |
|
"grad_norm": 0.5779871940612793, |
|
"learning_rate": 0.00015586713718131922, |
|
"loss": 0.4808, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.4906287069988138, |
|
"grad_norm": 0.5961576700210571, |
|
"learning_rate": 0.0001553443934520159, |
|
"loss": 0.4595, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.5001186239620403, |
|
"grad_norm": 0.5500461459159851, |
|
"learning_rate": 0.00015481946044447099, |
|
"loss": 0.4584, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.5096085409252669, |
|
"grad_norm": 0.7123764157295227, |
|
"learning_rate": 0.00015429235892364994, |
|
"loss": 0.4236, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.5190984578884934, |
|
"grad_norm": 0.6884598731994629, |
|
"learning_rate": 0.00015376310974029873, |
|
"loss": 0.4661, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.52858837485172, |
|
"grad_norm": 0.5527105331420898, |
|
"learning_rate": 0.0001532317338301192, |
|
"loss": 0.4307, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.5380782918149465, |
|
"grad_norm": 0.5720965266227722, |
|
"learning_rate": 0.00015269825221294098, |
|
"loss": 0.415, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.547568208778173, |
|
"grad_norm": 0.716526985168457, |
|
"learning_rate": 0.0001521626859918898, |
|
"loss": 0.48, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.5570581257413998, |
|
"grad_norm": 0.5414557456970215, |
|
"learning_rate": 0.00015162505635255287, |
|
"loss": 0.4943, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.5665480427046263, |
|
"grad_norm": 0.6315144896507263, |
|
"learning_rate": 0.0001510853845621409, |
|
"loss": 0.4427, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.5760379596678529, |
|
"grad_norm": 0.5746013522148132, |
|
"learning_rate": 0.00015054369196864644, |
|
"loss": 0.4148, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.5855278766310796, |
|
"grad_norm": 0.6795936226844788, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 0.4553, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.5950177935943062, |
|
"grad_norm": 0.6202004551887512, |
|
"learning_rate": 0.0001494543301632219, |
|
"loss": 0.4593, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.6045077105575327, |
|
"grad_norm": 0.49426236748695374, |
|
"learning_rate": 0.0001489067040435717, |
|
"loss": 0.4162, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.6139976275207593, |
|
"grad_norm": 0.6348617076873779, |
|
"learning_rate": 0.00014835714330369446, |
|
"loss": 0.4421, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.6234875444839858, |
|
"grad_norm": 0.5943213701248169, |
|
"learning_rate": 0.0001478056696827636, |
|
"loss": 0.491, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.6329774614472123, |
|
"grad_norm": 0.6268736124038696, |
|
"learning_rate": 0.00014725230499562119, |
|
"loss": 0.4144, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.6424673784104389, |
|
"grad_norm": 0.6147856712341309, |
|
"learning_rate": 0.00014669707113191483, |
|
"loss": 0.4089, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.6519572953736654, |
|
"grad_norm": 0.5978316068649292, |
|
"learning_rate": 0.00014613999005523174, |
|
"loss": 0.467, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.661447212336892, |
|
"grad_norm": 0.6331775784492493, |
|
"learning_rate": 0.00014558108380223012, |
|
"loss": 0.4489, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.6709371293001185, |
|
"grad_norm": 0.6398030519485474, |
|
"learning_rate": 0.00014502037448176734, |
|
"loss": 0.4383, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.680427046263345, |
|
"grad_norm": 0.6606591939926147, |
|
"learning_rate": 0.00014445788427402528, |
|
"loss": 0.4222, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.6899169632265718, |
|
"grad_norm": 0.644334077835083, |
|
"learning_rate": 0.00014389363542963306, |
|
"loss": 0.4403, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.6994068801897984, |
|
"grad_norm": 0.6094640493392944, |
|
"learning_rate": 0.00014332765026878687, |
|
"loss": 0.4433, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.708896797153025, |
|
"grad_norm": 0.6047070026397705, |
|
"learning_rate": 0.00014275995118036693, |
|
"loss": 0.4254, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.7183867141162514, |
|
"grad_norm": 0.612147331237793, |
|
"learning_rate": 0.00014219056062105193, |
|
"loss": 0.4544, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 1.7278766310794782, |
|
"grad_norm": 0.563838541507721, |
|
"learning_rate": 0.00014161950111443077, |
|
"loss": 0.4405, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.7373665480427047, |
|
"grad_norm": 0.6529393196105957, |
|
"learning_rate": 0.0001410467952501114, |
|
"loss": 0.4312, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.7468564650059313, |
|
"grad_norm": 0.663934588432312, |
|
"learning_rate": 0.00014047246568282736, |
|
"loss": 0.4421, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.7563463819691578, |
|
"grad_norm": 0.6207594871520996, |
|
"learning_rate": 0.00013989653513154165, |
|
"loss": 0.4365, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.7658362989323844, |
|
"grad_norm": 0.5731400847434998, |
|
"learning_rate": 0.0001393190263785479, |
|
"loss": 0.4202, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.775326215895611, |
|
"grad_norm": 0.5179402828216553, |
|
"learning_rate": 0.00013873996226856933, |
|
"loss": 0.4719, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 1.7848161328588374, |
|
"grad_norm": 0.40121665596961975, |
|
"learning_rate": 0.00013815936570785487, |
|
"loss": 0.4148, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.794306049822064, |
|
"grad_norm": 0.8414328694343567, |
|
"learning_rate": 0.00013757725966327322, |
|
"loss": 0.4648, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.8037959667852905, |
|
"grad_norm": 0.3349032998085022, |
|
"learning_rate": 0.00013699366716140435, |
|
"loss": 0.4168, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.813285883748517, |
|
"grad_norm": 0.6793477535247803, |
|
"learning_rate": 0.0001364086112876284, |
|
"loss": 0.4483, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 1.8227758007117436, |
|
"grad_norm": 0.5308493971824646, |
|
"learning_rate": 0.00013582211518521273, |
|
"loss": 0.4022, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.8322657176749704, |
|
"grad_norm": 0.5604270696640015, |
|
"learning_rate": 0.00013523420205439646, |
|
"loss": 0.4075, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 1.841755634638197, |
|
"grad_norm": 0.5878490805625916, |
|
"learning_rate": 0.00013464489515147238, |
|
"loss": 0.4337, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.8512455516014235, |
|
"grad_norm": 0.6031716465950012, |
|
"learning_rate": 0.00013405421778786737, |
|
"loss": 0.4655, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.8607354685646502, |
|
"grad_norm": 0.6215988993644714, |
|
"learning_rate": 0.00013346219332922016, |
|
"loss": 0.4715, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.8702253855278768, |
|
"grad_norm": 0.565021276473999, |
|
"learning_rate": 0.0001328688451944569, |
|
"loss": 0.4118, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 1.8797153024911033, |
|
"grad_norm": 0.5580743551254272, |
|
"learning_rate": 0.00013227419685486492, |
|
"loss": 0.4451, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.8892052194543298, |
|
"grad_norm": 0.7063333988189697, |
|
"learning_rate": 0.0001316782718331643, |
|
"loss": 0.4114, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 1.8986951364175564, |
|
"grad_norm": 0.534269392490387, |
|
"learning_rate": 0.00013108109370257712, |
|
"loss": 0.4029, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.8986951364175564, |
|
"eval_loss": 0.4455092251300812, |
|
"eval_runtime": 20.4087, |
|
"eval_samples_per_second": 15.533, |
|
"eval_steps_per_second": 7.791, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.908185053380783, |
|
"grad_norm": 0.5529626607894897, |
|
"learning_rate": 0.00013048268608589533, |
|
"loss": 0.4108, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 1.9176749703440095, |
|
"grad_norm": 0.673758327960968, |
|
"learning_rate": 0.00012988307265454597, |
|
"loss": 0.4219, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 1.927164887307236, |
|
"grad_norm": 0.6164252758026123, |
|
"learning_rate": 0.00012928227712765504, |
|
"loss": 0.4611, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 1.9366548042704625, |
|
"grad_norm": 0.6192930340766907, |
|
"learning_rate": 0.00012868032327110904, |
|
"loss": 0.4601, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.946144721233689, |
|
"grad_norm": 0.5321446061134338, |
|
"learning_rate": 0.00012807723489661495, |
|
"loss": 0.4112, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.9556346381969156, |
|
"grad_norm": 0.5794662237167358, |
|
"learning_rate": 0.0001274730358607583, |
|
"loss": 0.4915, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 1.9651245551601424, |
|
"grad_norm": 0.6252449750900269, |
|
"learning_rate": 0.00012686775006405946, |
|
"loss": 0.4307, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 1.974614472123369, |
|
"grad_norm": 0.6203203201293945, |
|
"learning_rate": 0.0001262614014500282, |
|
"loss": 0.4375, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 1.9841043890865955, |
|
"grad_norm": 0.617151141166687, |
|
"learning_rate": 0.00012565401400421651, |
|
"loss": 0.4517, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 1.993594306049822, |
|
"grad_norm": 0.681266188621521, |
|
"learning_rate": 0.00012504561175326985, |
|
"loss": 0.4245, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 2.0037959667852907, |
|
"grad_norm": 0.5491665005683899, |
|
"learning_rate": 0.0001244362187639767, |
|
"loss": 0.4221, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 2.0132858837485172, |
|
"grad_norm": 0.7722651958465576, |
|
"learning_rate": 0.0001238258591423165, |
|
"loss": 0.3264, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 2.022775800711744, |
|
"grad_norm": 0.6458817720413208, |
|
"learning_rate": 0.00012321455703250616, |
|
"loss": 0.3128, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 2.0322657176749703, |
|
"grad_norm": 0.5134835243225098, |
|
"learning_rate": 0.0001226023366160449, |
|
"loss": 0.3112, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 2.041755634638197, |
|
"grad_norm": 0.5023068785667419, |
|
"learning_rate": 0.00012198922211075778, |
|
"loss": 0.2929, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 2.0512455516014234, |
|
"grad_norm": 0.5612622499465942, |
|
"learning_rate": 0.00012137523776983757, |
|
"loss": 0.2943, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 2.06073546856465, |
|
"grad_norm": 0.5554709434509277, |
|
"learning_rate": 0.00012076040788088554, |
|
"loss": 0.3099, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 2.0702253855278765, |
|
"grad_norm": 0.5206819772720337, |
|
"learning_rate": 0.00012014475676495052, |
|
"loss": 0.2933, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 2.079715302491103, |
|
"grad_norm": 0.5239592790603638, |
|
"learning_rate": 0.000119528308775567, |
|
"loss": 0.2991, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 2.0892052194543296, |
|
"grad_norm": 0.5854765772819519, |
|
"learning_rate": 0.00011891108829779165, |
|
"loss": 0.3288, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.0986951364175566, |
|
"grad_norm": 0.7295845746994019, |
|
"learning_rate": 0.00011829311974723867, |
|
"loss": 0.3275, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 2.108185053380783, |
|
"grad_norm": 0.6106508374214172, |
|
"learning_rate": 0.00011767442756911417, |
|
"loss": 0.3111, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 2.1176749703440096, |
|
"grad_norm": 0.5574607849121094, |
|
"learning_rate": 0.00011705503623724898, |
|
"loss": 0.3194, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 2.127164887307236, |
|
"grad_norm": 0.5289790630340576, |
|
"learning_rate": 0.00011643497025313061, |
|
"loss": 0.3088, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 2.1366548042704627, |
|
"grad_norm": 0.5656944513320923, |
|
"learning_rate": 0.0001158142541449341, |
|
"loss": 0.3099, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 2.1461447212336893, |
|
"grad_norm": 0.594753086566925, |
|
"learning_rate": 0.0001151929124665516, |
|
"loss": 0.309, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 2.155634638196916, |
|
"grad_norm": 0.6289036870002747, |
|
"learning_rate": 0.00011457096979662114, |
|
"loss": 0.2948, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 2.1651245551601424, |
|
"grad_norm": 0.5498007535934448, |
|
"learning_rate": 0.00011394845073755455, |
|
"loss": 0.3082, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 2.174614472123369, |
|
"grad_norm": 0.5636598467826843, |
|
"learning_rate": 0.00011332537991456398, |
|
"loss": 0.2858, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 2.1841043890865954, |
|
"grad_norm": 0.6761645674705505, |
|
"learning_rate": 0.00011270178197468789, |
|
"loss": 0.332, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.193594306049822, |
|
"grad_norm": 0.5729044675827026, |
|
"learning_rate": 0.00011207768158581613, |
|
"loss": 0.326, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 2.2030842230130485, |
|
"grad_norm": 0.4295574724674225, |
|
"learning_rate": 0.00011145310343571411, |
|
"loss": 0.304, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 2.212574139976275, |
|
"grad_norm": 0.5968920588493347, |
|
"learning_rate": 0.0001108280722310462, |
|
"loss": 0.3205, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 2.2220640569395016, |
|
"grad_norm": 0.6211205720901489, |
|
"learning_rate": 0.00011020261269639842, |
|
"loss": 0.2823, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 2.231553973902728, |
|
"grad_norm": 0.6108773946762085, |
|
"learning_rate": 0.00010957674957330042, |
|
"loss": 0.324, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.241043890865955, |
|
"grad_norm": 0.8250358700752258, |
|
"learning_rate": 0.00010895050761924668, |
|
"loss": 0.3359, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 2.2505338078291817, |
|
"grad_norm": 0.5284693837165833, |
|
"learning_rate": 0.00010832391160671729, |
|
"loss": 0.3122, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 2.260023724792408, |
|
"grad_norm": 0.5513204336166382, |
|
"learning_rate": 0.00010769698632219794, |
|
"loss": 0.3264, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 2.2695136417556347, |
|
"grad_norm": 0.6842665672302246, |
|
"learning_rate": 0.00010706975656519946, |
|
"loss": 0.3047, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 2.2790035587188613, |
|
"grad_norm": 0.545037031173706, |
|
"learning_rate": 0.00010644224714727681, |
|
"loss": 0.3115, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.288493475682088, |
|
"grad_norm": 0.5622695088386536, |
|
"learning_rate": 0.00010581448289104758, |
|
"loss": 0.3158, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 2.2979833926453144, |
|
"grad_norm": 0.6832267642021179, |
|
"learning_rate": 0.00010518648862921012, |
|
"loss": 0.3298, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 2.307473309608541, |
|
"grad_norm": 0.5637592673301697, |
|
"learning_rate": 0.00010455828920356115, |
|
"loss": 0.3103, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 2.3169632265717675, |
|
"grad_norm": 0.5744633078575134, |
|
"learning_rate": 0.00010392990946401313, |
|
"loss": 0.3271, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 2.326453143534994, |
|
"grad_norm": 0.7118310928344727, |
|
"learning_rate": 0.00010330137426761135, |
|
"loss": 0.3276, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 2.3359430604982205, |
|
"grad_norm": 0.5940092206001282, |
|
"learning_rate": 0.00010267270847755048, |
|
"loss": 0.3152, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 2.345432977461447, |
|
"grad_norm": 0.6454032063484192, |
|
"learning_rate": 0.00010204393696219117, |
|
"loss": 0.2916, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 2.3549228944246736, |
|
"grad_norm": 0.5936978459358215, |
|
"learning_rate": 0.00010141508459407623, |
|
"loss": 0.3146, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 2.3644128113879006, |
|
"grad_norm": 0.5801250338554382, |
|
"learning_rate": 0.00010078617624894684, |
|
"loss": 0.3129, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 2.373902728351127, |
|
"grad_norm": 0.6835585832595825, |
|
"learning_rate": 0.00010015723680475846, |
|
"loss": 0.3028, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.373902728351127, |
|
"eval_loss": 0.4224609136581421, |
|
"eval_runtime": 20.4149, |
|
"eval_samples_per_second": 15.528, |
|
"eval_steps_per_second": 7.788, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.3833926453143537, |
|
"grad_norm": 0.5739105939865112, |
|
"learning_rate": 9.95282911406968e-05, |
|
"loss": 0.2934, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 2.3928825622775802, |
|
"grad_norm": 0.4810318350791931, |
|
"learning_rate": 9.889936413619356e-05, |
|
"loss": 0.3147, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 2.4023724792408068, |
|
"grad_norm": 0.5853947997093201, |
|
"learning_rate": 9.827048066994225e-05, |
|
"loss": 0.3008, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 2.4118623962040333, |
|
"grad_norm": 0.5773667693138123, |
|
"learning_rate": 9.764166561891432e-05, |
|
"loss": 0.2926, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 2.42135231316726, |
|
"grad_norm": 0.7114972472190857, |
|
"learning_rate": 9.70129438573747e-05, |
|
"loss": 0.3169, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.4308422301304864, |
|
"grad_norm": 0.6915512681007385, |
|
"learning_rate": 9.63843402558981e-05, |
|
"loss": 0.3068, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 2.440332147093713, |
|
"grad_norm": 0.594857394695282, |
|
"learning_rate": 9.57558796803852e-05, |
|
"loss": 0.3196, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 2.4498220640569395, |
|
"grad_norm": 0.6631921529769897, |
|
"learning_rate": 9.512758699107879e-05, |
|
"loss": 0.2999, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 2.459311981020166, |
|
"grad_norm": 0.6086694002151489, |
|
"learning_rate": 9.449948704158071e-05, |
|
"loss": 0.3128, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 2.4688018979833926, |
|
"grad_norm": 0.6541831493377686, |
|
"learning_rate": 9.38716046778684e-05, |
|
"loss": 0.3279, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.478291814946619, |
|
"grad_norm": 0.5784711241722107, |
|
"learning_rate": 9.324396473731217e-05, |
|
"loss": 0.3026, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 2.4877817319098456, |
|
"grad_norm": 0.6732935905456543, |
|
"learning_rate": 9.261659204769284e-05, |
|
"loss": 0.3212, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 2.497271648873072, |
|
"grad_norm": 0.6389648914337158, |
|
"learning_rate": 9.198951142621929e-05, |
|
"loss": 0.3184, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 2.5067615658362987, |
|
"grad_norm": 0.5656270980834961, |
|
"learning_rate": 9.136274767854716e-05, |
|
"loss": 0.31, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 2.5162514827995253, |
|
"grad_norm": 0.61983323097229, |
|
"learning_rate": 9.07363255977973e-05, |
|
"loss": 0.3148, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.525741399762752, |
|
"grad_norm": 0.46594876050949097, |
|
"learning_rate": 9.011026996357503e-05, |
|
"loss": 0.303, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 2.535231316725979, |
|
"grad_norm": 0.6862909197807312, |
|
"learning_rate": 8.948460554099018e-05, |
|
"loss": 0.3072, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 2.5447212336892053, |
|
"grad_norm": 0.6350931525230408, |
|
"learning_rate": 8.885935707967716e-05, |
|
"loss": 0.2909, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 2.554211150652432, |
|
"grad_norm": 0.6156574487686157, |
|
"learning_rate": 8.823454931281616e-05, |
|
"loss": 0.3375, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 2.5637010676156584, |
|
"grad_norm": 0.5949609875679016, |
|
"learning_rate": 8.76102069561545e-05, |
|
"loss": 0.3335, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.573190984578885, |
|
"grad_norm": 0.5457854866981506, |
|
"learning_rate": 8.698635470702923e-05, |
|
"loss": 0.3119, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 2.5826809015421115, |
|
"grad_norm": 0.5788692235946655, |
|
"learning_rate": 8.636301724339004e-05, |
|
"loss": 0.2898, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 2.592170818505338, |
|
"grad_norm": 0.5722967982292175, |
|
"learning_rate": 8.574021922282292e-05, |
|
"loss": 0.3079, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 2.6016607354685646, |
|
"grad_norm": 0.6501619815826416, |
|
"learning_rate": 8.511798528157512e-05, |
|
"loss": 0.2971, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 2.611150652431791, |
|
"grad_norm": 0.6138727068901062, |
|
"learning_rate": 8.449634003358022e-05, |
|
"loss": 0.3286, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.6206405693950177, |
|
"grad_norm": 0.5789212584495544, |
|
"learning_rate": 8.387530806948476e-05, |
|
"loss": 0.3101, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 2.630130486358244, |
|
"grad_norm": 0.6013932228088379, |
|
"learning_rate": 8.325491395567541e-05, |
|
"loss": 0.2997, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 2.639620403321471, |
|
"grad_norm": 0.5596510767936707, |
|
"learning_rate": 8.263518223330697e-05, |
|
"loss": 0.2928, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 2.6491103202846977, |
|
"grad_norm": 0.7271096706390381, |
|
"learning_rate": 8.201613741733203e-05, |
|
"loss": 0.3144, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 2.6586002372479243, |
|
"grad_norm": 0.715353786945343, |
|
"learning_rate": 8.13978039955308e-05, |
|
"loss": 0.3341, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.668090154211151, |
|
"grad_norm": 0.6036480665206909, |
|
"learning_rate": 8.078020642754274e-05, |
|
"loss": 0.3176, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 2.6775800711743774, |
|
"grad_norm": 0.5531415939331055, |
|
"learning_rate": 8.016336914389874e-05, |
|
"loss": 0.3043, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 2.687069988137604, |
|
"grad_norm": 0.5626965165138245, |
|
"learning_rate": 7.954731654505491e-05, |
|
"loss": 0.316, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 2.6965599051008304, |
|
"grad_norm": 0.6845198273658752, |
|
"learning_rate": 7.89320730004274e-05, |
|
"loss": 0.3167, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 2.706049822064057, |
|
"grad_norm": 0.5867395997047424, |
|
"learning_rate": 7.831766284742807e-05, |
|
"loss": 0.3189, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.7155397390272835, |
|
"grad_norm": 0.5502896308898926, |
|
"learning_rate": 7.77041103905023e-05, |
|
"loss": 0.3085, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 2.72502965599051, |
|
"grad_norm": 0.6936707496643066, |
|
"learning_rate": 7.709143990016702e-05, |
|
"loss": 0.2824, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 2.7345195729537366, |
|
"grad_norm": 0.6040688157081604, |
|
"learning_rate": 7.6479675612051e-05, |
|
"loss": 0.3005, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 2.744009489916963, |
|
"grad_norm": 0.6335172057151794, |
|
"learning_rate": 7.586884172593609e-05, |
|
"loss": 0.3048, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 2.7534994068801897, |
|
"grad_norm": 0.5545411705970764, |
|
"learning_rate": 7.525896240479976e-05, |
|
"loss": 0.3137, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.7629893238434162, |
|
"grad_norm": 0.56629878282547, |
|
"learning_rate": 7.465006177385953e-05, |
|
"loss": 0.3164, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 2.7724792408066428, |
|
"grad_norm": 0.6280866861343384, |
|
"learning_rate": 7.404216391961847e-05, |
|
"loss": 0.3266, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 2.7819691577698693, |
|
"grad_norm": 0.6720747947692871, |
|
"learning_rate": 7.343529288891239e-05, |
|
"loss": 0.2954, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 2.791459074733096, |
|
"grad_norm": 0.7422773838043213, |
|
"learning_rate": 7.282947268795877e-05, |
|
"loss": 0.3037, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 2.8009489916963224, |
|
"grad_norm": 0.6017013192176819, |
|
"learning_rate": 7.222472728140695e-05, |
|
"loss": 0.3007, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.8104389086595494, |
|
"grad_norm": 0.5282939672470093, |
|
"learning_rate": 7.162108059139032e-05, |
|
"loss": 0.2987, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 2.819928825622776, |
|
"grad_norm": 0.6391469240188599, |
|
"learning_rate": 7.101855649657991e-05, |
|
"loss": 0.306, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 2.8294187425860025, |
|
"grad_norm": 0.7014105319976807, |
|
"learning_rate": 7.041717883123977e-05, |
|
"loss": 0.2957, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 2.838908659549229, |
|
"grad_norm": 0.7407575845718384, |
|
"learning_rate": 6.981697138428434e-05, |
|
"loss": 0.2931, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 2.8483985765124555, |
|
"grad_norm": 0.6662490367889404, |
|
"learning_rate": 6.921795789833723e-05, |
|
"loss": 0.2909, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.8483985765124555, |
|
"eval_loss": 0.3944380581378937, |
|
"eval_runtime": 20.3916, |
|
"eval_samples_per_second": 15.546, |
|
"eval_steps_per_second": 7.797, |
|
"step": 3000 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 5000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.9630137686215885e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|