{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 8415, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0023767082590612004, "grad_norm": 3.921875, "learning_rate": 2.3767082590612003e-06, "loss": 3.6948, "step": 20 }, { "epoch": 0.004753416518122401, "grad_norm": 5.28125, "learning_rate": 4.753416518122401e-06, "loss": 3.8478, "step": 40 }, { "epoch": 0.0071301247771836, "grad_norm": 7.21875, "learning_rate": 7.130124777183601e-06, "loss": 3.6042, "step": 60 }, { "epoch": 0.009506833036244802, "grad_norm": 2.625, "learning_rate": 9.506833036244801e-06, "loss": 3.5338, "step": 80 }, { "epoch": 0.011883541295306001, "grad_norm": 1.515625, "learning_rate": 1.1883541295306002e-05, "loss": 3.2504, "step": 100 }, { "epoch": 0.0142602495543672, "grad_norm": 1.3203125, "learning_rate": 1.4260249554367203e-05, "loss": 2.956, "step": 120 }, { "epoch": 0.016636957813428402, "grad_norm": 1.6015625, "learning_rate": 1.66369578134284e-05, "loss": 2.9073, "step": 140 }, { "epoch": 0.019013666072489603, "grad_norm": 1.0703125, "learning_rate": 1.9013666072489602e-05, "loss": 2.6436, "step": 160 }, { "epoch": 0.0213903743315508, "grad_norm": 1.0234375, "learning_rate": 2.1390374331550803e-05, "loss": 2.6018, "step": 180 }, { "epoch": 0.023767082590612002, "grad_norm": 1.0234375, "learning_rate": 2.3767082590612004e-05, "loss": 2.4289, "step": 200 }, { "epoch": 0.026143790849673203, "grad_norm": 1.109375, "learning_rate": 2.6143790849673204e-05, "loss": 2.5239, "step": 220 }, { "epoch": 0.0285204991087344, "grad_norm": 1.5625, "learning_rate": 2.8520499108734405e-05, "loss": 2.4617, "step": 240 }, { "epoch": 0.030897207367795602, "grad_norm": 1.1171875, "learning_rate": 3.08972073677956e-05, "loss": 2.396, "step": 260 }, { "epoch": 0.033273915626856804, "grad_norm": 1.03125, "learning_rate": 3.32739156268568e-05, "loss": 2.2682, "step": 280 }, { "epoch": 0.035650623885918005, "grad_norm": 1.265625, "learning_rate": 3.5650623885918004e-05, "loss": 2.3627, "step": 300 }, { "epoch": 0.038027332144979206, "grad_norm": 1.1484375, "learning_rate": 3.8027332144979205e-05, "loss": 2.3072, "step": 320 }, { "epoch": 0.04040404040404041, "grad_norm": 1.5859375, "learning_rate": 4.0404040404040405e-05, "loss": 2.2124, "step": 340 }, { "epoch": 0.0427807486631016, "grad_norm": 1.0703125, "learning_rate": 4.2780748663101606e-05, "loss": 2.2619, "step": 360 }, { "epoch": 0.0451574569221628, "grad_norm": 1.46875, "learning_rate": 4.515745692216281e-05, "loss": 2.1928, "step": 380 }, { "epoch": 0.047534165181224004, "grad_norm": 1.3515625, "learning_rate": 4.753416518122401e-05, "loss": 2.0428, "step": 400 }, { "epoch": 0.049910873440285206, "grad_norm": 1.90625, "learning_rate": 4.991087344028521e-05, "loss": 1.9929, "step": 420 }, { "epoch": 0.05228758169934641, "grad_norm": 1.3828125, "learning_rate": 5.228758169934641e-05, "loss": 2.146, "step": 440 }, { "epoch": 0.05466428995840761, "grad_norm": 1.5859375, "learning_rate": 5.466428995840761e-05, "loss": 2.0289, "step": 460 }, { "epoch": 0.0570409982174688, "grad_norm": 1.3125, "learning_rate": 5.704099821746881e-05, "loss": 1.9554, "step": 480 }, { "epoch": 0.059417706476530004, "grad_norm": 1.6484375, "learning_rate": 5.941770647653001e-05, "loss": 1.9871, "step": 500 }, { "epoch": 0.061794414735591205, "grad_norm": 1.9140625, "learning_rate": 6.17944147355912e-05, "loss": 2.0099, "step": 520 }, { "epoch": 0.06417112299465241, "grad_norm": 2.140625, "learning_rate": 6.41711229946524e-05, "loss": 1.9199, "step": 540 }, { "epoch": 0.06654783125371361, "grad_norm": 2.203125, "learning_rate": 6.65478312537136e-05, "loss": 2.0074, "step": 560 }, { "epoch": 0.0689245395127748, "grad_norm": 1.2421875, "learning_rate": 6.892453951277481e-05, "loss": 1.8052, "step": 580 }, { "epoch": 0.07130124777183601, "grad_norm": 2.25, "learning_rate": 7.130124777183601e-05, "loss": 2.0366, "step": 600 }, { "epoch": 0.0736779560308972, "grad_norm": 2.4375, "learning_rate": 7.367795603089721e-05, "loss": 1.8426, "step": 620 }, { "epoch": 0.07605466428995841, "grad_norm": 3.75, "learning_rate": 7.605466428995841e-05, "loss": 1.6836, "step": 640 }, { "epoch": 0.0784313725490196, "grad_norm": 3.09375, "learning_rate": 7.843137254901961e-05, "loss": 2.0483, "step": 660 }, { "epoch": 0.08080808080808081, "grad_norm": 1.6953125, "learning_rate": 8.080808080808081e-05, "loss": 1.902, "step": 680 }, { "epoch": 0.08318478906714201, "grad_norm": 2.109375, "learning_rate": 8.318478906714201e-05, "loss": 1.9887, "step": 700 }, { "epoch": 0.0855614973262032, "grad_norm": 2.984375, "learning_rate": 8.556149732620321e-05, "loss": 1.6542, "step": 720 }, { "epoch": 0.08793820558526441, "grad_norm": 3.3125, "learning_rate": 8.793820558526441e-05, "loss": 1.6718, "step": 740 }, { "epoch": 0.0903149138443256, "grad_norm": 2.9375, "learning_rate": 9.031491384432561e-05, "loss": 1.706, "step": 760 }, { "epoch": 0.09269162210338681, "grad_norm": 3.59375, "learning_rate": 9.269162210338681e-05, "loss": 1.629, "step": 780 }, { "epoch": 0.09506833036244801, "grad_norm": 2.40625, "learning_rate": 9.506833036244802e-05, "loss": 1.6748, "step": 800 }, { "epoch": 0.0974450386215092, "grad_norm": 2.578125, "learning_rate": 9.744503862150922e-05, "loss": 1.8269, "step": 820 }, { "epoch": 0.09982174688057041, "grad_norm": 2.328125, "learning_rate": 9.982174688057042e-05, "loss": 1.8727, "step": 840 }, { "epoch": 0.1021984551396316, "grad_norm": 1.8671875, "learning_rate": 0.00010219845513963162, "loss": 1.7455, "step": 860 }, { "epoch": 0.10457516339869281, "grad_norm": 2.140625, "learning_rate": 0.00010457516339869282, "loss": 1.7395, "step": 880 }, { "epoch": 0.10695187165775401, "grad_norm": 1.609375, "learning_rate": 0.00010695187165775402, "loss": 1.7935, "step": 900 }, { "epoch": 0.10932857991681522, "grad_norm": 1.6796875, "learning_rate": 0.00010932857991681522, "loss": 1.8009, "step": 920 }, { "epoch": 0.11170528817587641, "grad_norm": 1.8515625, "learning_rate": 0.00011170528817587642, "loss": 1.5582, "step": 940 }, { "epoch": 0.1140819964349376, "grad_norm": 2.0, "learning_rate": 0.00011408199643493762, "loss": 1.6402, "step": 960 }, { "epoch": 0.11645870469399881, "grad_norm": 1.9921875, "learning_rate": 0.00011645870469399882, "loss": 1.665, "step": 980 }, { "epoch": 0.11883541295306001, "grad_norm": 1.78125, "learning_rate": 0.00011883541295306002, "loss": 1.8407, "step": 1000 }, { "epoch": 0.12121212121212122, "grad_norm": 1.9296875, "learning_rate": 0.00012121212121212122, "loss": 1.7665, "step": 1020 }, { "epoch": 0.12358882947118241, "grad_norm": 2.5625, "learning_rate": 0.0001235888294711824, "loss": 1.8593, "step": 1040 }, { "epoch": 0.1259655377302436, "grad_norm": 2.109375, "learning_rate": 0.00012596553773024362, "loss": 1.7738, "step": 1060 }, { "epoch": 0.12834224598930483, "grad_norm": 2.078125, "learning_rate": 0.0001283422459893048, "loss": 1.8227, "step": 1080 }, { "epoch": 0.13071895424836602, "grad_norm": 3.53125, "learning_rate": 0.00013071895424836603, "loss": 1.7159, "step": 1100 }, { "epoch": 0.13309566250742721, "grad_norm": 1.890625, "learning_rate": 0.0001330956625074272, "loss": 1.6953, "step": 1120 }, { "epoch": 0.1354723707664884, "grad_norm": 1.8671875, "learning_rate": 0.00013547237076648843, "loss": 1.5567, "step": 1140 }, { "epoch": 0.1378490790255496, "grad_norm": 1.515625, "learning_rate": 0.00013784907902554961, "loss": 1.7637, "step": 1160 }, { "epoch": 0.14022578728461083, "grad_norm": 2.375, "learning_rate": 0.00014022578728461083, "loss": 1.5884, "step": 1180 }, { "epoch": 0.14260249554367202, "grad_norm": 2.046875, "learning_rate": 0.00014260249554367202, "loss": 1.5542, "step": 1200 }, { "epoch": 0.14497920380273321, "grad_norm": 2.515625, "learning_rate": 0.00014497920380273323, "loss": 1.5124, "step": 1220 }, { "epoch": 0.1473559120617944, "grad_norm": 1.609375, "learning_rate": 0.00014735591206179442, "loss": 1.5822, "step": 1240 }, { "epoch": 0.1497326203208556, "grad_norm": 1.7890625, "learning_rate": 0.00014973262032085563, "loss": 1.6216, "step": 1260 }, { "epoch": 0.15210932857991682, "grad_norm": 2.34375, "learning_rate": 0.00015210932857991682, "loss": 1.7366, "step": 1280 }, { "epoch": 0.15448603683897802, "grad_norm": 1.8828125, "learning_rate": 0.00015448603683897803, "loss": 1.3673, "step": 1300 }, { "epoch": 0.1568627450980392, "grad_norm": 1.8046875, "learning_rate": 0.00015686274509803922, "loss": 1.4071, "step": 1320 }, { "epoch": 0.1592394533571004, "grad_norm": 2.59375, "learning_rate": 0.00015923945335710043, "loss": 1.6515, "step": 1340 }, { "epoch": 0.16161616161616163, "grad_norm": 2.28125, "learning_rate": 0.00016161616161616162, "loss": 1.7604, "step": 1360 }, { "epoch": 0.16399286987522282, "grad_norm": 1.9140625, "learning_rate": 0.0001639928698752228, "loss": 1.5925, "step": 1380 }, { "epoch": 0.16636957813428402, "grad_norm": 3.40625, "learning_rate": 0.00016636957813428402, "loss": 1.5874, "step": 1400 }, { "epoch": 0.1687462863933452, "grad_norm": 2.65625, "learning_rate": 0.0001687462863933452, "loss": 1.6116, "step": 1420 }, { "epoch": 0.1711229946524064, "grad_norm": 1.8671875, "learning_rate": 0.00017112299465240642, "loss": 1.544, "step": 1440 }, { "epoch": 0.17349970291146763, "grad_norm": 2.15625, "learning_rate": 0.0001734997029114676, "loss": 1.5015, "step": 1460 }, { "epoch": 0.17587641117052882, "grad_norm": 1.8515625, "learning_rate": 0.00017587641117052883, "loss": 1.5858, "step": 1480 }, { "epoch": 0.17825311942959002, "grad_norm": 1.90625, "learning_rate": 0.00017825311942959, "loss": 1.5469, "step": 1500 }, { "epoch": 0.1806298276886512, "grad_norm": 2.421875, "learning_rate": 0.00018062982768865123, "loss": 1.5092, "step": 1520 }, { "epoch": 0.1830065359477124, "grad_norm": 1.9921875, "learning_rate": 0.00018300653594771241, "loss": 1.3362, "step": 1540 }, { "epoch": 0.18538324420677363, "grad_norm": 1.6015625, "learning_rate": 0.00018538324420677363, "loss": 1.5088, "step": 1560 }, { "epoch": 0.18775995246583482, "grad_norm": 1.9765625, "learning_rate": 0.00018775995246583482, "loss": 1.5518, "step": 1580 }, { "epoch": 0.19013666072489602, "grad_norm": 1.8046875, "learning_rate": 0.00019013666072489603, "loss": 1.4344, "step": 1600 }, { "epoch": 0.1925133689839572, "grad_norm": 1.71875, "learning_rate": 0.00019251336898395722, "loss": 1.5411, "step": 1620 }, { "epoch": 0.1948900772430184, "grad_norm": 1.515625, "learning_rate": 0.00019489007724301843, "loss": 1.5832, "step": 1640 }, { "epoch": 0.19726678550207963, "grad_norm": 1.71875, "learning_rate": 0.00019726678550207962, "loss": 1.5468, "step": 1660 }, { "epoch": 0.19964349376114082, "grad_norm": 2.078125, "learning_rate": 0.00019964349376114083, "loss": 1.3053, "step": 1680 }, { "epoch": 0.20202020202020202, "grad_norm": 2.0625, "learning_rate": 0.0001999968531423333, "loss": 1.5303, "step": 1700 }, { "epoch": 0.2043969102792632, "grad_norm": 1.4921875, "learning_rate": 0.0001999850935511802, "loss": 1.6331, "step": 1720 }, { "epoch": 0.20677361853832443, "grad_norm": 1.8203125, "learning_rate": 0.0001999646242513938, "loss": 1.5621, "step": 1740 }, { "epoch": 0.20915032679738563, "grad_norm": 1.578125, "learning_rate": 0.00019993544702605638, "loss": 1.4932, "step": 1760 }, { "epoch": 0.21152703505644682, "grad_norm": 2.09375, "learning_rate": 0.00019989756441679795, "loss": 1.3936, "step": 1780 }, { "epoch": 0.21390374331550802, "grad_norm": 1.4453125, "learning_rate": 0.00019985097972357547, "loss": 1.3227, "step": 1800 }, { "epoch": 0.2162804515745692, "grad_norm": 1.671875, "learning_rate": 0.0001997956970043848, "loss": 1.4253, "step": 1820 }, { "epoch": 0.21865715983363043, "grad_norm": 1.5078125, "learning_rate": 0.00019973172107490758, "loss": 1.3946, "step": 1840 }, { "epoch": 0.22103386809269163, "grad_norm": 1.625, "learning_rate": 0.00019965905750809158, "loss": 1.2779, "step": 1860 }, { "epoch": 0.22341057635175282, "grad_norm": 1.4296875, "learning_rate": 0.00019957771263366526, "loss": 1.5051, "step": 1880 }, { "epoch": 0.22578728461081402, "grad_norm": 1.9765625, "learning_rate": 0.00019948769353758653, "loss": 1.3954, "step": 1900 }, { "epoch": 0.2281639928698752, "grad_norm": 2.25, "learning_rate": 0.00019938900806142518, "loss": 1.2049, "step": 1920 }, { "epoch": 0.23054070112893643, "grad_norm": 2.5625, "learning_rate": 0.00019928166480168011, "loss": 1.5335, "step": 1940 }, { "epoch": 0.23291740938799763, "grad_norm": 2.03125, "learning_rate": 0.00019916567310903035, "loss": 1.3868, "step": 1960 }, { "epoch": 0.23529411764705882, "grad_norm": 1.46875, "learning_rate": 0.0001990410430875205, "loss": 1.3677, "step": 1980 }, { "epoch": 0.23767082590612001, "grad_norm": 2.484375, "learning_rate": 0.0001989077855936807, "loss": 1.2757, "step": 2000 }, { "epoch": 0.24004753416518124, "grad_norm": 2.21875, "learning_rate": 0.0001987659122355806, "loss": 1.4893, "step": 2020 }, { "epoch": 0.24242424242424243, "grad_norm": 1.28125, "learning_rate": 0.00019861543537181867, "loss": 1.2123, "step": 2040 }, { "epoch": 0.24480095068330363, "grad_norm": 2.078125, "learning_rate": 0.00019845636811044521, "loss": 1.4714, "step": 2060 }, { "epoch": 0.24717765894236482, "grad_norm": 2.453125, "learning_rate": 0.0001982887243078207, "loss": 1.463, "step": 2080 }, { "epoch": 0.24955436720142601, "grad_norm": 1.953125, "learning_rate": 0.00019811251856740873, "loss": 1.5086, "step": 2100 }, { "epoch": 0.2519310754604872, "grad_norm": 1.734375, "learning_rate": 0.000197927766238504, "loss": 1.3604, "step": 2120 }, { "epoch": 0.25430778371954843, "grad_norm": 1.96875, "learning_rate": 0.00019773448341489495, "loss": 1.2226, "step": 2140 }, { "epoch": 0.25668449197860965, "grad_norm": 1.609375, "learning_rate": 0.00019753268693346225, "loss": 1.2003, "step": 2160 }, { "epoch": 0.2590612002376708, "grad_norm": 1.9296875, "learning_rate": 0.0001973223943727117, "loss": 1.192, "step": 2180 }, { "epoch": 0.26143790849673204, "grad_norm": 1.9765625, "learning_rate": 0.00019710362405124334, "loss": 1.2616, "step": 2200 }, { "epoch": 0.2638146167557932, "grad_norm": 1.2265625, "learning_rate": 0.0001968763950261554, "loss": 1.2493, "step": 2220 }, { "epoch": 0.26619132501485443, "grad_norm": 2.234375, "learning_rate": 0.0001966407270913846, "loss": 1.278, "step": 2240 }, { "epoch": 0.26856803327391565, "grad_norm": 1.5703125, "learning_rate": 0.00019639664077598142, "loss": 1.3797, "step": 2260 }, { "epoch": 0.2709447415329768, "grad_norm": 2.078125, "learning_rate": 0.0001961441573423223, "loss": 1.1243, "step": 2280 }, { "epoch": 0.27332144979203804, "grad_norm": 2.0, "learning_rate": 0.0001958832987842571, "loss": 1.324, "step": 2300 }, { "epoch": 0.2756981580510992, "grad_norm": 2.375, "learning_rate": 0.00019561408782519345, "loss": 1.3572, "step": 2320 }, { "epoch": 0.27807486631016043, "grad_norm": 2.265625, "learning_rate": 0.0001953365479161172, "loss": 1.3404, "step": 2340 }, { "epoch": 0.28045157456922165, "grad_norm": 2.984375, "learning_rate": 0.00019505070323354965, "loss": 1.3237, "step": 2360 }, { "epoch": 0.2828282828282828, "grad_norm": 1.8203125, "learning_rate": 0.0001947565786774415, "loss": 1.2894, "step": 2380 }, { "epoch": 0.28520499108734404, "grad_norm": 1.7578125, "learning_rate": 0.0001944541998690038, "loss": 1.3445, "step": 2400 }, { "epoch": 0.2875816993464052, "grad_norm": 2.015625, "learning_rate": 0.0001941435931484761, "loss": 1.3859, "step": 2420 }, { "epoch": 0.28995840760546643, "grad_norm": 1.7890625, "learning_rate": 0.00019382478557283204, "loss": 1.2185, "step": 2440 }, { "epoch": 0.29233511586452765, "grad_norm": 1.546875, "learning_rate": 0.00019349780491342223, "loss": 1.3545, "step": 2460 }, { "epoch": 0.2947118241235888, "grad_norm": 2.125, "learning_rate": 0.00019316267965355528, "loss": 1.2694, "step": 2480 }, { "epoch": 0.29708853238265004, "grad_norm": 2.03125, "learning_rate": 0.00019281943898601645, "loss": 1.0126, "step": 2500 }, { "epoch": 0.2994652406417112, "grad_norm": 1.6328125, "learning_rate": 0.00019246811281052487, "loss": 1.2593, "step": 2520 }, { "epoch": 0.3018419489007724, "grad_norm": 1.578125, "learning_rate": 0.00019210873173112865, "loss": 1.3609, "step": 2540 }, { "epoch": 0.30421865715983365, "grad_norm": 1.375, "learning_rate": 0.0001917413270535393, "loss": 0.9368, "step": 2560 }, { "epoch": 0.3065953654188948, "grad_norm": 2.25, "learning_rate": 0.0001913659307824045, "loss": 1.1893, "step": 2580 }, { "epoch": 0.30897207367795604, "grad_norm": 2.21875, "learning_rate": 0.0001909825756185202, "loss": 1.1345, "step": 2600 }, { "epoch": 0.3113487819370172, "grad_norm": 1.515625, "learning_rate": 0.0001905912949559821, "loss": 1.2925, "step": 2620 }, { "epoch": 0.3137254901960784, "grad_norm": 1.296875, "learning_rate": 0.00019019212287927663, "loss": 1.0842, "step": 2640 }, { "epoch": 0.31610219845513965, "grad_norm": 2.40625, "learning_rate": 0.00018978509416031186, "loss": 1.2722, "step": 2660 }, { "epoch": 0.3184789067142008, "grad_norm": 1.796875, "learning_rate": 0.00018937024425538855, "loss": 1.3566, "step": 2680 }, { "epoch": 0.32085561497326204, "grad_norm": 2.09375, "learning_rate": 0.0001889476093021115, "loss": 0.9899, "step": 2700 }, { "epoch": 0.32323232323232326, "grad_norm": 1.8046875, "learning_rate": 0.00018851722611624164, "loss": 1.0713, "step": 2720 }, { "epoch": 0.3256090314913844, "grad_norm": 2.25, "learning_rate": 0.00018807913218848906, "loss": 1.1957, "step": 2740 }, { "epoch": 0.32798573975044565, "grad_norm": 1.453125, "learning_rate": 0.000187633365681247, "loss": 1.3258, "step": 2760 }, { "epoch": 0.3303624480095068, "grad_norm": 1.5625, "learning_rate": 0.00018717996542526777, "loss": 1.1146, "step": 2780 }, { "epoch": 0.33273915626856804, "grad_norm": 1.984375, "learning_rate": 0.00018671897091627993, "loss": 1.0215, "step": 2800 }, { "epoch": 0.33511586452762926, "grad_norm": 1.5703125, "learning_rate": 0.00018625042231154817, "loss": 1.1107, "step": 2820 }, { "epoch": 0.3374925727866904, "grad_norm": 1.8828125, "learning_rate": 0.00018577436042637477, "loss": 1.1655, "step": 2840 }, { "epoch": 0.33986928104575165, "grad_norm": 1.765625, "learning_rate": 0.00018529082673054457, "loss": 1.1197, "step": 2860 }, { "epoch": 0.3422459893048128, "grad_norm": 1.8828125, "learning_rate": 0.0001847998633447123, "loss": 1.1792, "step": 2880 }, { "epoch": 0.34462269756387404, "grad_norm": 1.9765625, "learning_rate": 0.0001843015130367335, "loss": 1.1933, "step": 2900 }, { "epoch": 0.34699940582293526, "grad_norm": 2.15625, "learning_rate": 0.00018379581921793914, "loss": 1.1603, "step": 2920 }, { "epoch": 0.3493761140819964, "grad_norm": 1.09375, "learning_rate": 0.00018328282593935377, "loss": 1.1653, "step": 2940 }, { "epoch": 0.35175282234105765, "grad_norm": 1.7578125, "learning_rate": 0.00018276257788785855, "loss": 1.0807, "step": 2960 }, { "epoch": 0.3541295306001188, "grad_norm": 2.296875, "learning_rate": 0.00018223512038229833, "loss": 1.0923, "step": 2980 }, { "epoch": 0.35650623885918004, "grad_norm": 1.7109375, "learning_rate": 0.00018170049936953406, "loss": 1.3022, "step": 3000 }, { "epoch": 0.35888294711824126, "grad_norm": 1.625, "learning_rate": 0.00018115876142044032, "loss": 1.2415, "step": 3020 }, { "epoch": 0.3612596553773024, "grad_norm": 1.4453125, "learning_rate": 0.0001806099537258485, "loss": 1.0849, "step": 3040 }, { "epoch": 0.36363636363636365, "grad_norm": 1.8203125, "learning_rate": 0.00018005412409243606, "loss": 1.1712, "step": 3060 }, { "epoch": 0.3660130718954248, "grad_norm": 1.4921875, "learning_rate": 0.000179491320938562, "loss": 1.2795, "step": 3080 }, { "epoch": 0.36838978015448604, "grad_norm": 1.734375, "learning_rate": 0.00017892159329004916, "loss": 1.1304, "step": 3100 }, { "epoch": 0.37076648841354726, "grad_norm": 2.265625, "learning_rate": 0.00017834499077591374, "loss": 1.277, "step": 3120 }, { "epoch": 0.3731431966726084, "grad_norm": 1.5546875, "learning_rate": 0.00017776156362404186, "loss": 1.2222, "step": 3140 }, { "epoch": 0.37551990493166965, "grad_norm": 2.4375, "learning_rate": 0.0001771713626568143, "loss": 1.2639, "step": 3160 }, { "epoch": 0.3778966131907308, "grad_norm": 1.7109375, "learning_rate": 0.0001765744392866795, "loss": 1.1679, "step": 3180 }, { "epoch": 0.38027332144979203, "grad_norm": 1.9921875, "learning_rate": 0.00017597084551167476, "loss": 1.0657, "step": 3200 }, { "epoch": 0.38265002970885326, "grad_norm": 1.2109375, "learning_rate": 0.00017536063391089697, "loss": 1.1886, "step": 3220 }, { "epoch": 0.3850267379679144, "grad_norm": 1.1875, "learning_rate": 0.00017474385763992212, "loss": 1.1526, "step": 3240 }, { "epoch": 0.38740344622697565, "grad_norm": 1.6171875, "learning_rate": 0.00017412057042617525, "loss": 1.0648, "step": 3260 }, { "epoch": 0.3897801544860368, "grad_norm": 1.8515625, "learning_rate": 0.00017349082656424995, "loss": 1.2938, "step": 3280 }, { "epoch": 0.39215686274509803, "grad_norm": 1.34375, "learning_rate": 0.00017285468091117904, "loss": 1.1246, "step": 3300 }, { "epoch": 0.39453357100415926, "grad_norm": 2.03125, "learning_rate": 0.00017221218888165572, "loss": 1.0179, "step": 3320 }, { "epoch": 0.3969102792632204, "grad_norm": 1.8671875, "learning_rate": 0.0001715634064432065, "loss": 0.914, "step": 3340 }, { "epoch": 0.39928698752228164, "grad_norm": 1.4296875, "learning_rate": 0.0001709083901113159, "loss": 1.1192, "step": 3360 }, { "epoch": 0.40166369578134287, "grad_norm": 3.203125, "learning_rate": 0.00017024719694450337, "loss": 1.1462, "step": 3380 }, { "epoch": 0.40404040404040403, "grad_norm": 1.3515625, "learning_rate": 0.00016957988453935276, "loss": 0.9452, "step": 3400 }, { "epoch": 0.40641711229946526, "grad_norm": 2.28125, "learning_rate": 0.00016890651102549538, "loss": 1.1837, "step": 3420 }, { "epoch": 0.4087938205585264, "grad_norm": 2.515625, "learning_rate": 0.00016822713506054604, "loss": 0.8889, "step": 3440 }, { "epoch": 0.41117052881758764, "grad_norm": 1.515625, "learning_rate": 0.0001675418158249935, "loss": 1.0873, "step": 3460 }, { "epoch": 0.41354723707664887, "grad_norm": 1.3515625, "learning_rate": 0.0001668506130170453, "loss": 0.995, "step": 3480 }, { "epoch": 0.41592394533571003, "grad_norm": 1.7421875, "learning_rate": 0.0001661535868474273, "loss": 1.1595, "step": 3500 }, { "epoch": 0.41830065359477125, "grad_norm": 2.15625, "learning_rate": 0.00016545079803413892, "loss": 1.1134, "step": 3520 }, { "epoch": 0.4206773618538324, "grad_norm": 0.92578125, "learning_rate": 0.00016474230779716384, "loss": 0.7802, "step": 3540 }, { "epoch": 0.42305407011289364, "grad_norm": 1.671875, "learning_rate": 0.00016402817785313712, "loss": 0.9085, "step": 3560 }, { "epoch": 0.42543077837195487, "grad_norm": 2.390625, "learning_rate": 0.00016330847040996915, "loss": 1.0042, "step": 3580 }, { "epoch": 0.42780748663101603, "grad_norm": 2.46875, "learning_rate": 0.00016258324816142668, "loss": 1.0318, "step": 3600 }, { "epoch": 0.43018419489007725, "grad_norm": 3.375, "learning_rate": 0.00016185257428167143, "loss": 1.0833, "step": 3620 }, { "epoch": 0.4325609031491384, "grad_norm": 2.078125, "learning_rate": 0.00016111651241975734, "loss": 1.0208, "step": 3640 }, { "epoch": 0.43493761140819964, "grad_norm": 2.125, "learning_rate": 0.00016037512669408565, "loss": 1.0074, "step": 3660 }, { "epoch": 0.43731431966726086, "grad_norm": 1.5703125, "learning_rate": 0.0001596284816868198, "loss": 1.2403, "step": 3680 }, { "epoch": 0.43969102792632203, "grad_norm": 2.59375, "learning_rate": 0.00015887664243825967, "loss": 1.0694, "step": 3700 }, { "epoch": 0.44206773618538325, "grad_norm": 1.3203125, "learning_rate": 0.0001581196744411759, "loss": 1.0033, "step": 3720 }, { "epoch": 0.4444444444444444, "grad_norm": 2.359375, "learning_rate": 0.0001573576436351046, "loss": 1.0574, "step": 3740 }, { "epoch": 0.44682115270350564, "grad_norm": 2.109375, "learning_rate": 0.00015659061640060378, "loss": 1.2204, "step": 3760 }, { "epoch": 0.44919786096256686, "grad_norm": 1.8046875, "learning_rate": 0.0001558186595534705, "loss": 1.157, "step": 3780 }, { "epoch": 0.45157456922162803, "grad_norm": 1.6171875, "learning_rate": 0.0001550418403389208, "loss": 0.9542, "step": 3800 }, { "epoch": 0.45395127748068925, "grad_norm": 2.390625, "learning_rate": 0.00015426022642573193, "loss": 0.9281, "step": 3820 }, { "epoch": 0.4563279857397504, "grad_norm": 1.4453125, "learning_rate": 0.00015347388590034757, "loss": 0.9331, "step": 3840 }, { "epoch": 0.45870469399881164, "grad_norm": 1.3828125, "learning_rate": 0.00015268288726094705, "loss": 0.7978, "step": 3860 }, { "epoch": 0.46108140225787286, "grad_norm": 2.421875, "learning_rate": 0.00015188729941147824, "loss": 0.9807, "step": 3880 }, { "epoch": 0.46345811051693403, "grad_norm": 2.640625, "learning_rate": 0.0001510871916556555, "loss": 1.0974, "step": 3900 }, { "epoch": 0.46583481877599525, "grad_norm": 1.171875, "learning_rate": 0.00015028263369092253, "loss": 0.9807, "step": 3920 }, { "epoch": 0.4682115270350565, "grad_norm": 1.6484375, "learning_rate": 0.00014947369560238104, "loss": 1.0588, "step": 3940 }, { "epoch": 0.47058823529411764, "grad_norm": 2.546875, "learning_rate": 0.00014866044785668563, "loss": 1.0442, "step": 3960 }, { "epoch": 0.47296494355317886, "grad_norm": 2.9375, "learning_rate": 0.00014784296129590548, "loss": 0.8837, "step": 3980 }, { "epoch": 0.47534165181224003, "grad_norm": 1.7890625, "learning_rate": 0.00014702130713135317, "loss": 1.129, "step": 4000 }, { "epoch": 0.47771836007130125, "grad_norm": 2.171875, "learning_rate": 0.00014619555693738166, "loss": 1.0496, "step": 4020 }, { "epoch": 0.4800950683303625, "grad_norm": 2.328125, "learning_rate": 0.0001453657826451493, "loss": 1.0136, "step": 4040 }, { "epoch": 0.48247177658942364, "grad_norm": 2.203125, "learning_rate": 0.00014453205653635376, "loss": 0.9794, "step": 4060 }, { "epoch": 0.48484848484848486, "grad_norm": 2.734375, "learning_rate": 0.00014369445123693596, "loss": 0.7651, "step": 4080 }, { "epoch": 0.48722519310754603, "grad_norm": 2.15625, "learning_rate": 0.0001428530397107533, "loss": 0.9667, "step": 4100 }, { "epoch": 0.48960190136660725, "grad_norm": 2.578125, "learning_rate": 0.0001420078952532238, "loss": 0.8897, "step": 4120 }, { "epoch": 0.4919786096256685, "grad_norm": 4.28125, "learning_rate": 0.0001411590914849415, "loss": 1.0208, "step": 4140 }, { "epoch": 0.49435531788472964, "grad_norm": 2.34375, "learning_rate": 0.00014030670234526323, "loss": 1.0138, "step": 4160 }, { "epoch": 0.49673202614379086, "grad_norm": 1.234375, "learning_rate": 0.00013945080208586775, "loss": 0.8435, "step": 4180 }, { "epoch": 0.49910873440285203, "grad_norm": 2.828125, "learning_rate": 0.0001385914652642877, "loss": 0.7405, "step": 4200 }, { "epoch": 0.5014854426619133, "grad_norm": 2.59375, "learning_rate": 0.00013772876673741498, "loss": 0.8999, "step": 4220 }, { "epoch": 0.5038621509209744, "grad_norm": 2.171875, "learning_rate": 0.00013686278165497977, "loss": 0.9063, "step": 4240 }, { "epoch": 0.5062388591800356, "grad_norm": 3.578125, "learning_rate": 0.00013599358545300438, "loss": 0.8485, "step": 4260 }, { "epoch": 0.5086155674390969, "grad_norm": 2.5625, "learning_rate": 0.00013512125384723204, "loss": 0.7843, "step": 4280 }, { "epoch": 0.5109922756981581, "grad_norm": 1.5546875, "learning_rate": 0.00013424586282653116, "loss": 0.7548, "step": 4300 }, { "epoch": 0.5133689839572193, "grad_norm": 2.8125, "learning_rate": 0.00013336748864627593, "loss": 0.9609, "step": 4320 }, { "epoch": 0.5157456922162804, "grad_norm": 2.453125, "learning_rate": 0.00013248620782170396, "loss": 0.9358, "step": 4340 }, { "epoch": 0.5181224004753416, "grad_norm": 2.21875, "learning_rate": 0.00013160209712125074, "loss": 1.0639, "step": 4360 }, { "epoch": 0.5204991087344029, "grad_norm": 1.515625, "learning_rate": 0.0001307152335598624, "loss": 1.0147, "step": 4380 }, { "epoch": 0.5228758169934641, "grad_norm": 2.4375, "learning_rate": 0.00012982569439228713, "loss": 0.7749, "step": 4400 }, { "epoch": 0.5252525252525253, "grad_norm": 2.21875, "learning_rate": 0.0001289335571063453, "loss": 0.642, "step": 4420 }, { "epoch": 0.5276292335115864, "grad_norm": 2.984375, "learning_rate": 0.00012803889941617944, "loss": 0.863, "step": 4440 }, { "epoch": 0.5300059417706476, "grad_norm": 2.265625, "learning_rate": 0.0001271417992554849, "loss": 0.8817, "step": 4460 }, { "epoch": 0.5323826500297089, "grad_norm": 2.3125, "learning_rate": 0.00012624233477072057, "loss": 0.9788, "step": 4480 }, { "epoch": 0.5347593582887701, "grad_norm": 2.484375, "learning_rate": 0.00012534058431430198, "loss": 0.8273, "step": 4500 }, { "epoch": 0.5371360665478313, "grad_norm": 1.2265625, "learning_rate": 0.0001244366264377757, "loss": 0.848, "step": 4520 }, { "epoch": 0.5395127748068924, "grad_norm": 2.625, "learning_rate": 0.00012353053988497684, "loss": 0.8556, "step": 4540 }, { "epoch": 0.5418894830659536, "grad_norm": 2.375, "learning_rate": 0.00012262240358516967, "loss": 0.8137, "step": 4560 }, { "epoch": 0.5442661913250149, "grad_norm": 1.1015625, "learning_rate": 0.00012171229664617208, "loss": 0.7362, "step": 4580 }, { "epoch": 0.5466428995840761, "grad_norm": 2.90625, "learning_rate": 0.00012080029834746447, "loss": 0.7504, "step": 4600 }, { "epoch": 0.5490196078431373, "grad_norm": 3.109375, "learning_rate": 0.00011988648813328367, "loss": 1.0486, "step": 4620 }, { "epoch": 0.5513963161021984, "grad_norm": 1.875, "learning_rate": 0.00011897094560570265, "loss": 0.7487, "step": 4640 }, { "epoch": 0.5537730243612596, "grad_norm": 2.171875, "learning_rate": 0.00011805375051769636, "loss": 0.9325, "step": 4660 }, { "epoch": 0.5561497326203209, "grad_norm": 2.28125, "learning_rate": 0.00011713498276619432, "loss": 0.906, "step": 4680 }, { "epoch": 0.5585264408793821, "grad_norm": 1.953125, "learning_rate": 0.0001162147223851209, "loss": 0.7214, "step": 4700 }, { "epoch": 0.5609031491384433, "grad_norm": 2.34375, "learning_rate": 0.0001152930495384236, "loss": 1.01, "step": 4720 }, { "epoch": 0.5632798573975044, "grad_norm": 2.53125, "learning_rate": 0.00011437004451308983, "loss": 1.0897, "step": 4740 }, { "epoch": 0.5656565656565656, "grad_norm": 3.296875, "learning_rate": 0.00011344578771215319, "loss": 0.8222, "step": 4760 }, { "epoch": 0.5680332739156269, "grad_norm": 2.921875, "learning_rate": 0.00011252035964768961, "loss": 0.8411, "step": 4780 }, { "epoch": 0.5704099821746881, "grad_norm": 1.9375, "learning_rate": 0.00011159384093380377, "loss": 0.747, "step": 4800 }, { "epoch": 0.5727866904337493, "grad_norm": 2.375, "learning_rate": 0.00011066631227960693, "loss": 0.7352, "step": 4820 }, { "epoch": 0.5751633986928104, "grad_norm": 2.484375, "learning_rate": 0.00010973785448218639, "loss": 0.901, "step": 4840 }, { "epoch": 0.5775401069518716, "grad_norm": 1.7265625, "learning_rate": 0.00010880854841956712, "loss": 0.6489, "step": 4860 }, { "epoch": 0.5799168152109329, "grad_norm": 1.3359375, "learning_rate": 0.00010787847504366649, "loss": 0.7034, "step": 4880 }, { "epoch": 0.5822935234699941, "grad_norm": 3.046875, "learning_rate": 0.00010694771537324269, "loss": 0.7427, "step": 4900 }, { "epoch": 0.5846702317290553, "grad_norm": 3.40625, "learning_rate": 0.00010601635048683698, "loss": 0.7655, "step": 4920 }, { "epoch": 0.5870469399881164, "grad_norm": 2.21875, "learning_rate": 0.00010508446151571109, "loss": 0.7467, "step": 4940 }, { "epoch": 0.5894236482471776, "grad_norm": 1.9375, "learning_rate": 0.0001041521296367798, "loss": 0.7942, "step": 4960 }, { "epoch": 0.5918003565062389, "grad_norm": 3.34375, "learning_rate": 0.00010321943606553961, "loss": 1.033, "step": 4980 }, { "epoch": 0.5941770647653001, "grad_norm": 2.765625, "learning_rate": 0.00010228646204899401, "loss": 0.7049, "step": 5000 }, { "epoch": 0.5965537730243613, "grad_norm": 1.59375, "learning_rate": 0.0001013532888585762, "loss": 0.7016, "step": 5020 }, { "epoch": 0.5989304812834224, "grad_norm": 2.875, "learning_rate": 0.00010041999778306936, "loss": 0.9511, "step": 5040 }, { "epoch": 0.6013071895424836, "grad_norm": 3.640625, "learning_rate": 9.948667012152566e-05, "loss": 0.6895, "step": 5060 }, { "epoch": 0.6036838978015449, "grad_norm": 3.3125, "learning_rate": 9.855338717618432e-05, "loss": 0.8696, "step": 5080 }, { "epoch": 0.6060606060606061, "grad_norm": 2.625, "learning_rate": 9.762023024538926e-05, "loss": 0.8621, "step": 5100 }, { "epoch": 0.6084373143196673, "grad_norm": 1.984375, "learning_rate": 9.668728061650733e-05, "loss": 0.9548, "step": 5120 }, { "epoch": 0.6108140225787284, "grad_norm": 3.40625, "learning_rate": 9.575461955884726e-05, "loss": 0.8048, "step": 5140 }, { "epoch": 0.6131907308377896, "grad_norm": 2.265625, "learning_rate": 9.482232831658034e-05, "loss": 0.8866, "step": 5160 }, { "epoch": 0.6155674390968509, "grad_norm": 1.5390625, "learning_rate": 9.389048810166317e-05, "loss": 0.7404, "step": 5180 }, { "epoch": 0.6179441473559121, "grad_norm": 1.359375, "learning_rate": 9.29591800867634e-05, "loss": 0.6817, "step": 5200 }, { "epoch": 0.6203208556149733, "grad_norm": 5.0, "learning_rate": 9.202848539818865e-05, "loss": 0.7504, "step": 5220 }, { "epoch": 0.6226975638740344, "grad_norm": 2.484375, "learning_rate": 9.109848510881961e-05, "loss": 0.7767, "step": 5240 }, { "epoch": 0.6250742721330956, "grad_norm": 2.53125, "learning_rate": 9.016926023104789e-05, "loss": 0.7846, "step": 5260 }, { "epoch": 0.6274509803921569, "grad_norm": 3.421875, "learning_rate": 8.924089170971887e-05, "loss": 0.8327, "step": 5280 }, { "epoch": 0.6298276886512181, "grad_norm": 3.765625, "learning_rate": 8.831346041508069e-05, "loss": 0.8035, "step": 5300 }, { "epoch": 0.6322043969102793, "grad_norm": 2.25, "learning_rate": 8.738704713573959e-05, "loss": 0.9569, "step": 5320 }, { "epoch": 0.6345811051693404, "grad_norm": 2.34375, "learning_rate": 8.646173257162245e-05, "loss": 0.6629, "step": 5340 }, { "epoch": 0.6369578134284016, "grad_norm": 2.875, "learning_rate": 8.553759732694696e-05, "loss": 0.6597, "step": 5360 }, { "epoch": 0.6393345216874629, "grad_norm": 2.859375, "learning_rate": 8.461472190320021e-05, "loss": 0.6857, "step": 5380 }, { "epoch": 0.6417112299465241, "grad_norm": 1.5234375, "learning_rate": 8.369318669212625e-05, "loss": 0.7234, "step": 5400 }, { "epoch": 0.6440879382055853, "grad_norm": 3.359375, "learning_rate": 8.277307196872303e-05, "loss": 0.7483, "step": 5420 }, { "epoch": 0.6464646464646465, "grad_norm": 1.21875, "learning_rate": 8.185445788424974e-05, "loss": 0.7538, "step": 5440 }, { "epoch": 0.6488413547237076, "grad_norm": 3.234375, "learning_rate": 8.093742445924491e-05, "loss": 0.7882, "step": 5460 }, { "epoch": 0.6512180629827689, "grad_norm": 3.546875, "learning_rate": 8.002205157655554e-05, "loss": 0.6984, "step": 5480 }, { "epoch": 0.6535947712418301, "grad_norm": 2.28125, "learning_rate": 7.910841897437875e-05, "loss": 0.7044, "step": 5500 }, { "epoch": 0.6559714795008913, "grad_norm": 3.390625, "learning_rate": 7.819660623931575e-05, "loss": 0.6569, "step": 5520 }, { "epoch": 0.6583481877599525, "grad_norm": 1.5859375, "learning_rate": 7.728669279943897e-05, "loss": 0.8069, "step": 5540 }, { "epoch": 0.6607248960190136, "grad_norm": 3.125, "learning_rate": 7.637875791737299e-05, "loss": 0.759, "step": 5560 }, { "epoch": 0.6631016042780749, "grad_norm": 1.4921875, "learning_rate": 7.547288068339025e-05, "loss": 0.6806, "step": 5580 }, { "epoch": 0.6654783125371361, "grad_norm": 2.875, "learning_rate": 7.456914000852123e-05, "loss": 0.7777, "step": 5600 }, { "epoch": 0.6678550207961973, "grad_norm": 1.453125, "learning_rate": 7.366761461768052e-05, "loss": 0.7375, "step": 5620 }, { "epoch": 0.6702317290552585, "grad_norm": 1.734375, "learning_rate": 7.276838304280935e-05, "loss": 0.6862, "step": 5640 }, { "epoch": 0.6726084373143196, "grad_norm": 3.171875, "learning_rate": 7.187152361603432e-05, "loss": 0.6195, "step": 5660 }, { "epoch": 0.6749851455733809, "grad_norm": 1.8046875, "learning_rate": 7.097711446284405e-05, "loss": 0.5486, "step": 5680 }, { "epoch": 0.6773618538324421, "grad_norm": 2.203125, "learning_rate": 7.008523349528377e-05, "loss": 0.7452, "step": 5700 }, { "epoch": 0.6797385620915033, "grad_norm": 3.03125, "learning_rate": 6.919595840516815e-05, "loss": 0.7417, "step": 5720 }, { "epoch": 0.6821152703505645, "grad_norm": 3.1875, "learning_rate": 6.830936665731371e-05, "loss": 0.6798, "step": 5740 }, { "epoch": 0.6844919786096256, "grad_norm": 2.328125, "learning_rate": 6.742553548279095e-05, "loss": 0.7222, "step": 5760 }, { "epoch": 0.6868686868686869, "grad_norm": 1.0625, "learning_rate": 6.654454187219649e-05, "loss": 0.6629, "step": 5780 }, { "epoch": 0.6892453951277481, "grad_norm": 1.6015625, "learning_rate": 6.56664625689466e-05, "loss": 0.7808, "step": 5800 }, { "epoch": 0.6916221033868093, "grad_norm": 2.421875, "learning_rate": 6.479137406259206e-05, "loss": 0.7898, "step": 5820 }, { "epoch": 0.6939988116458705, "grad_norm": 1.53125, "learning_rate": 6.39193525821551e-05, "loss": 0.8595, "step": 5840 }, { "epoch": 0.6963755199049316, "grad_norm": 3.21875, "learning_rate": 6.305047408948908e-05, "loss": 0.8169, "step": 5860 }, { "epoch": 0.6987522281639929, "grad_norm": 2.796875, "learning_rate": 6.21848142726615e-05, "loss": 0.7892, "step": 5880 }, { "epoch": 0.7011289364230541, "grad_norm": 2.671875, "learning_rate": 6.13224485393608e-05, "loss": 0.8179, "step": 5900 }, { "epoch": 0.7035056446821153, "grad_norm": 1.8046875, "learning_rate": 6.046345201032748e-05, "loss": 0.6881, "step": 5920 }, { "epoch": 0.7058823529411765, "grad_norm": 1.515625, "learning_rate": 5.960789951281052e-05, "loss": 0.9263, "step": 5940 }, { "epoch": 0.7082590612002376, "grad_norm": 2.21875, "learning_rate": 5.8755865574049016e-05, "loss": 0.7661, "step": 5960 }, { "epoch": 0.7106357694592988, "grad_norm": 2.5625, "learning_rate": 5.7907424414780135e-05, "loss": 0.7932, "step": 5980 }, { "epoch": 0.7130124777183601, "grad_norm": 1.8984375, "learning_rate": 5.706264994277386e-05, "loss": 0.8045, "step": 6000 }, { "epoch": 0.7153891859774213, "grad_norm": 1.09375, "learning_rate": 5.6221615746394644e-05, "loss": 0.6996, "step": 6020 }, { "epoch": 0.7177658942364825, "grad_norm": 1.796875, "learning_rate": 5.538439508819139e-05, "loss": 0.7662, "step": 6040 }, { "epoch": 0.7201426024955436, "grad_norm": 1.890625, "learning_rate": 5.4551060898515404e-05, "loss": 0.6049, "step": 6060 }, { "epoch": 0.7225193107546048, "grad_norm": 2.15625, "learning_rate": 5.372168576916732e-05, "loss": 0.6502, "step": 6080 }, { "epoch": 0.7248960190136661, "grad_norm": 2.6875, "learning_rate": 5.289634194707387e-05, "loss": 0.6785, "step": 6100 }, { "epoch": 0.7272727272727273, "grad_norm": 2.890625, "learning_rate": 5.207510132799436e-05, "loss": 0.7902, "step": 6120 }, { "epoch": 0.7296494355317885, "grad_norm": 2.703125, "learning_rate": 5.125803545025758e-05, "loss": 0.6948, "step": 6140 }, { "epoch": 0.7320261437908496, "grad_norm": 2.84375, "learning_rate": 5.0445215488530525e-05, "loss": 0.7113, "step": 6160 }, { "epoch": 0.7344028520499108, "grad_norm": 3.484375, "learning_rate": 4.963671224761808e-05, "loss": 0.6249, "step": 6180 }, { "epoch": 0.7367795603089721, "grad_norm": 1.59375, "learning_rate": 4.883259615629515e-05, "loss": 0.7295, "step": 6200 }, { "epoch": 0.7391562685680333, "grad_norm": 1.359375, "learning_rate": 4.8032937261171896e-05, "loss": 0.8041, "step": 6220 }, { "epoch": 0.7415329768270945, "grad_norm": 1.8046875, "learning_rate": 4.7237805220591744e-05, "loss": 0.6296, "step": 6240 }, { "epoch": 0.7439096850861556, "grad_norm": 1.609375, "learning_rate": 4.644726929856342e-05, "loss": 0.8887, "step": 6260 }, { "epoch": 0.7462863933452168, "grad_norm": 2.421875, "learning_rate": 4.5661398358727524e-05, "loss": 0.6501, "step": 6280 }, { "epoch": 0.7486631016042781, "grad_norm": 1.7109375, "learning_rate": 4.4880260858357746e-05, "loss": 0.5981, "step": 6300 }, { "epoch": 0.7510398098633393, "grad_norm": 1.765625, "learning_rate": 4.4103924842397395e-05, "loss": 0.5993, "step": 6320 }, { "epoch": 0.7534165181224005, "grad_norm": 3.484375, "learning_rate": 4.3332457937532246e-05, "loss": 0.8187, "step": 6340 }, { "epoch": 0.7557932263814616, "grad_norm": 7.09375, "learning_rate": 4.256592734629947e-05, "loss": 0.826, "step": 6360 }, { "epoch": 0.7581699346405228, "grad_norm": 1.578125, "learning_rate": 4.18043998412335e-05, "loss": 0.5091, "step": 6380 }, { "epoch": 0.7605466428995841, "grad_norm": 0.78125, "learning_rate": 4.104794175904966e-05, "loss": 0.609, "step": 6400 }, { "epoch": 0.7629233511586453, "grad_norm": 3.171875, "learning_rate": 4.02966189948655e-05, "loss": 0.7031, "step": 6420 }, { "epoch": 0.7653000594177065, "grad_norm": 1.6015625, "learning_rate": 3.955049699646054e-05, "loss": 0.5454, "step": 6440 }, { "epoch": 0.7676767676767676, "grad_norm": 3.578125, "learning_rate": 3.880964075857535e-05, "loss": 0.629, "step": 6460 }, { "epoch": 0.7700534759358288, "grad_norm": 2.609375, "learning_rate": 3.80741148172497e-05, "loss": 0.6376, "step": 6480 }, { "epoch": 0.7724301841948901, "grad_norm": 2.015625, "learning_rate": 3.734398324420073e-05, "loss": 0.6744, "step": 6500 }, { "epoch": 0.7748068924539513, "grad_norm": 3.140625, "learning_rate": 3.661930964124193e-05, "loss": 0.6133, "step": 6520 }, { "epoch": 0.7771836007130125, "grad_norm": 1.734375, "learning_rate": 3.5900157134742574e-05, "loss": 0.5723, "step": 6540 }, { "epoch": 0.7795603089720736, "grad_norm": 3.921875, "learning_rate": 3.5186588370128746e-05, "loss": 0.6248, "step": 6560 }, { "epoch": 0.7819370172311348, "grad_norm": 3.203125, "learning_rate": 3.447866550642649e-05, "loss": 0.6323, "step": 6580 }, { "epoch": 0.7843137254901961, "grad_norm": 3.1875, "learning_rate": 3.377645021084701e-05, "loss": 0.7705, "step": 6600 }, { "epoch": 0.7866904337492573, "grad_norm": 2.546875, "learning_rate": 3.3080003653414724e-05, "loss": 0.5787, "step": 6620 }, { "epoch": 0.7890671420083185, "grad_norm": 1.2109375, "learning_rate": 3.238938650163899e-05, "loss": 0.7117, "step": 6640 }, { "epoch": 0.7914438502673797, "grad_norm": 1.640625, "learning_rate": 3.17046589152292e-05, "loss": 0.5979, "step": 6660 }, { "epoch": 0.7938205585264408, "grad_norm": 4.53125, "learning_rate": 3.10258805408542e-05, "loss": 0.6202, "step": 6680 }, { "epoch": 0.7961972667855021, "grad_norm": 3.484375, "learning_rate": 3.0353110506946647e-05, "loss": 0.6462, "step": 6700 }, { "epoch": 0.7985739750445633, "grad_norm": 2.640625, "learning_rate": 2.968640741855223e-05, "loss": 0.576, "step": 6720 }, { "epoch": 0.8009506833036245, "grad_norm": 1.875, "learning_rate": 2.9025829352224477e-05, "loss": 0.7645, "step": 6740 }, { "epoch": 0.8033273915626857, "grad_norm": 2.453125, "learning_rate": 2.8371433850965922e-05, "loss": 0.6348, "step": 6760 }, { "epoch": 0.8057040998217468, "grad_norm": 2.875, "learning_rate": 2.7723277919215397e-05, "loss": 0.7165, "step": 6780 }, { "epoch": 0.8080808080808081, "grad_norm": 1.7890625, "learning_rate": 2.70814180178823e-05, "loss": 0.6772, "step": 6800 }, { "epoch": 0.8104575163398693, "grad_norm": 3.25, "learning_rate": 2.644591005942846e-05, "loss": 0.6995, "step": 6820 }, { "epoch": 0.8128342245989305, "grad_norm": 3.546875, "learning_rate": 2.5816809402997522e-05, "loss": 0.6519, "step": 6840 }, { "epoch": 0.8152109328579917, "grad_norm": 4.375, "learning_rate": 2.5194170849592492e-05, "loss": 0.612, "step": 6860 }, { "epoch": 0.8175876411170528, "grad_norm": 3.40625, "learning_rate": 2.4578048637302208e-05, "loss": 0.8347, "step": 6880 }, { "epoch": 0.8199643493761141, "grad_norm": 3.953125, "learning_rate": 2.396849643657657e-05, "loss": 0.7846, "step": 6900 }, { "epoch": 0.8223410576351753, "grad_norm": 3.0, "learning_rate": 2.3365567345551233e-05, "loss": 0.6799, "step": 6920 }, { "epoch": 0.8247177658942365, "grad_norm": 2.15625, "learning_rate": 2.276931388542235e-05, "loss": 0.704, "step": 6940 }, { "epoch": 0.8270944741532977, "grad_norm": 3.3125, "learning_rate": 2.2179787995871403e-05, "loss": 0.5832, "step": 6960 }, { "epoch": 0.8294711824123588, "grad_norm": 1.8046875, "learning_rate": 2.1597041030540643e-05, "loss": 0.6464, "step": 6980 }, { "epoch": 0.8318478906714201, "grad_norm": 1.0859375, "learning_rate": 2.1021123752559836e-05, "loss": 0.6596, "step": 7000 }, { "epoch": 0.8342245989304813, "grad_norm": 2.4375, "learning_rate": 2.0452086330124164e-05, "loss": 0.6124, "step": 7020 }, { "epoch": 0.8366013071895425, "grad_norm": 1.953125, "learning_rate": 1.988997833212406e-05, "loss": 0.8721, "step": 7040 }, { "epoch": 0.8389780154486037, "grad_norm": 1.0234375, "learning_rate": 1.933484872382737e-05, "loss": 0.6729, "step": 7060 }, { "epoch": 0.8413547237076648, "grad_norm": 3.859375, "learning_rate": 1.8786745862613885e-05, "loss": 0.7747, "step": 7080 }, { "epoch": 0.8437314319667261, "grad_norm": 2.703125, "learning_rate": 1.8245717493762925e-05, "loss": 0.6807, "step": 7100 }, { "epoch": 0.8461081402257873, "grad_norm": 2.4375, "learning_rate": 1.7711810746294312e-05, "loss": 0.851, "step": 7120 }, { "epoch": 0.8484848484848485, "grad_norm": 1.4765625, "learning_rate": 1.7185072128862933e-05, "loss": 0.6449, "step": 7140 }, { "epoch": 0.8508615567439097, "grad_norm": 2.140625, "learning_rate": 1.6665547525707316e-05, "loss": 0.6683, "step": 7160 }, { "epoch": 0.8532382650029708, "grad_norm": 2.25, "learning_rate": 1.6153282192652698e-05, "loss": 0.5818, "step": 7180 }, { "epoch": 0.8556149732620321, "grad_norm": 2.1875, "learning_rate": 1.5648320753168844e-05, "loss": 0.6614, "step": 7200 }, { "epoch": 0.8579916815210933, "grad_norm": 2.5625, "learning_rate": 1.5150707194482695e-05, "loss": 0.7086, "step": 7220 }, { "epoch": 0.8603683897801545, "grad_norm": 2.734375, "learning_rate": 1.4660484863746938e-05, "loss": 0.778, "step": 7240 }, { "epoch": 0.8627450980392157, "grad_norm": 2.109375, "learning_rate": 1.4177696464263723e-05, "loss": 0.7629, "step": 7260 }, { "epoch": 0.8651218062982768, "grad_norm": 3.1875, "learning_rate": 1.3702384051765005e-05, "loss": 0.8923, "step": 7280 }, { "epoch": 0.8674985145573381, "grad_norm": 4.125, "learning_rate": 1.3234589030748956e-05, "loss": 0.6005, "step": 7300 }, { "epoch": 0.8698752228163993, "grad_norm": 2.9375, "learning_rate": 1.2774352150873203e-05, "loss": 0.741, "step": 7320 }, { "epoch": 0.8722519310754605, "grad_norm": 2.78125, "learning_rate": 1.2321713503405208e-05, "loss": 0.5276, "step": 7340 }, { "epoch": 0.8746286393345217, "grad_norm": 3.28125, "learning_rate": 1.187671251772987e-05, "loss": 0.5997, "step": 7360 }, { "epoch": 0.8770053475935828, "grad_norm": 1.8125, "learning_rate": 1.143938795791476e-05, "loss": 0.6653, "step": 7380 }, { "epoch": 0.8793820558526441, "grad_norm": 2.46875, "learning_rate": 1.1009777919333507e-05, "loss": 0.5658, "step": 7400 }, { "epoch": 0.8817587641117053, "grad_norm": 2.609375, "learning_rate": 1.0587919825347236e-05, "loss": 0.6695, "step": 7420 }, { "epoch": 0.8841354723707665, "grad_norm": 1.7734375, "learning_rate": 1.0173850424044596e-05, "loss": 0.68, "step": 7440 }, { "epoch": 0.8865121806298277, "grad_norm": 3.640625, "learning_rate": 9.76760578504068e-06, "loss": 0.8137, "step": 7460 }, { "epoch": 0.8888888888888888, "grad_norm": 1.8125, "learning_rate": 9.369221296335006e-06, "loss": 0.7129, "step": 7480 }, { "epoch": 0.8912655971479501, "grad_norm": 4.625, "learning_rate": 8.978731661228768e-06, "loss": 0.6628, "step": 7500 }, { "epoch": 0.8936423054070113, "grad_norm": 4.125, "learning_rate": 8.596170895301959e-06, "loss": 0.6621, "step": 7520 }, { "epoch": 0.8960190136660725, "grad_norm": 2.953125, "learning_rate": 8.221572323450222e-06, "loss": 0.6171, "step": 7540 }, { "epoch": 0.8983957219251337, "grad_norm": 3.078125, "learning_rate": 7.854968576981824e-06, "loss": 0.5361, "step": 7560 }, { "epoch": 0.9007724301841948, "grad_norm": 3.8125, "learning_rate": 7.49639159077532e-06, "loss": 0.6402, "step": 7580 }, { "epoch": 0.9031491384432561, "grad_norm": 3.25, "learning_rate": 7.145872600497561e-06, "loss": 0.5515, "step": 7600 }, { "epoch": 0.9055258467023173, "grad_norm": 2.078125, "learning_rate": 6.8034421398827765e-06, "loss": 0.6165, "step": 7620 }, { "epoch": 0.9079025549613785, "grad_norm": 3.78125, "learning_rate": 6.469130038072835e-06, "loss": 0.8606, "step": 7640 }, { "epoch": 0.9102792632204397, "grad_norm": 2.875, "learning_rate": 6.142965417018798e-06, "loss": 0.6364, "step": 7660 }, { "epoch": 0.9126559714795008, "grad_norm": 2.53125, "learning_rate": 5.824976688944051e-06, "loss": 0.578, "step": 7680 }, { "epoch": 0.9150326797385621, "grad_norm": 2.15625, "learning_rate": 5.515191553869381e-06, "loss": 0.5698, "step": 7700 }, { "epoch": 0.9174093879976233, "grad_norm": 2.96875, "learning_rate": 5.213636997200044e-06, "loss": 0.6748, "step": 7720 }, { "epoch": 0.9197860962566845, "grad_norm": 4.0625, "learning_rate": 4.920339287374942e-06, "loss": 0.7089, "step": 7740 }, { "epoch": 0.9221628045157457, "grad_norm": 3.578125, "learning_rate": 4.635323973578543e-06, "loss": 0.7648, "step": 7760 }, { "epoch": 0.9245395127748068, "grad_norm": 1.0390625, "learning_rate": 4.3586158835151495e-06, "loss": 0.6778, "step": 7780 }, { "epoch": 0.9269162210338681, "grad_norm": 3.578125, "learning_rate": 4.090239121246231e-06, "loss": 0.7377, "step": 7800 }, { "epoch": 0.9292929292929293, "grad_norm": 3.5, "learning_rate": 3.830217065090702e-06, "loss": 0.6488, "step": 7820 }, { "epoch": 0.9316696375519905, "grad_norm": 1.40625, "learning_rate": 3.5785723655884287e-06, "loss": 0.5778, "step": 7840 }, { "epoch": 0.9340463458110517, "grad_norm": 3.3125, "learning_rate": 3.335326943527117e-06, "loss": 0.6642, "step": 7860 }, { "epoch": 0.936423054070113, "grad_norm": 3.421875, "learning_rate": 3.100501988032878e-06, "loss": 0.5973, "step": 7880 }, { "epoch": 0.9387997623291741, "grad_norm": 4.71875, "learning_rate": 2.874117954724309e-06, "loss": 0.618, "step": 7900 }, { "epoch": 0.9411764705882353, "grad_norm": 3.921875, "learning_rate": 2.656194563930714e-06, "loss": 0.7053, "step": 7920 }, { "epoch": 0.9435531788472965, "grad_norm": 2.953125, "learning_rate": 2.446750798974229e-06, "loss": 0.6074, "step": 7940 }, { "epoch": 0.9459298871063577, "grad_norm": 2.46875, "learning_rate": 2.2458049045161244e-06, "loss": 0.5246, "step": 7960 }, { "epoch": 0.948306595365419, "grad_norm": 3.484375, "learning_rate": 2.0533743849676436e-06, "loss": 0.648, "step": 7980 }, { "epoch": 0.9506833036244801, "grad_norm": 2.953125, "learning_rate": 1.869476002965065e-06, "loss": 0.6625, "step": 8000 }, { "epoch": 0.9530600118835413, "grad_norm": 1.5859375, "learning_rate": 1.69412577790955e-06, "loss": 0.7165, "step": 8020 }, { "epoch": 0.9554367201426025, "grad_norm": 1.3046875, "learning_rate": 1.5273389845717245e-06, "loss": 0.6024, "step": 8040 }, { "epoch": 0.9578134284016637, "grad_norm": 3.6875, "learning_rate": 1.3691301517610554e-06, "loss": 0.6466, "step": 8060 }, { "epoch": 0.960190136660725, "grad_norm": 1.171875, "learning_rate": 1.2195130610602623e-06, "loss": 0.6671, "step": 8080 }, { "epoch": 0.9625668449197861, "grad_norm": 1.046875, "learning_rate": 1.0785007456247886e-06, "loss": 0.628, "step": 8100 }, { "epoch": 0.9649435531788473, "grad_norm": 2.9375, "learning_rate": 9.461054890474996e-07, "loss": 0.776, "step": 8120 }, { "epoch": 0.9673202614379085, "grad_norm": 1.671875, "learning_rate": 8.223388242886265e-07, "loss": 0.6368, "step": 8140 }, { "epoch": 0.9696969696969697, "grad_norm": 2.625, "learning_rate": 7.072115326711704e-07, "loss": 0.5938, "step": 8160 }, { "epoch": 0.972073677956031, "grad_norm": 1.171875, "learning_rate": 6.00733642941742e-07, "loss": 0.6331, "step": 8180 }, { "epoch": 0.9744503862150921, "grad_norm": 2.234375, "learning_rate": 5.029144303968724e-07, "loss": 0.566, "step": 8200 }, { "epoch": 0.9768270944741533, "grad_norm": 3.5, "learning_rate": 4.1376241607518074e-07, "loss": 0.6529, "step": 8220 }, { "epoch": 0.9792038027332145, "grad_norm": 0.91015625, "learning_rate": 3.332853660149904e-07, "loss": 0.7165, "step": 8240 }, { "epoch": 0.9815805109922757, "grad_norm": 2.5625, "learning_rate": 2.6149029057785936e-07, "loss": 0.6302, "step": 8260 }, { "epoch": 0.983957219251337, "grad_norm": 0.97265625, "learning_rate": 1.9838344383793505e-07, "loss": 0.6525, "step": 8280 }, { "epoch": 0.9863339275103981, "grad_norm": 2.90625, "learning_rate": 1.4397032303715697e-07, "loss": 0.6024, "step": 8300 }, { "epoch": 0.9887106357694593, "grad_norm": 2.421875, "learning_rate": 9.825566810633958e-08, "loss": 0.4994, "step": 8320 }, { "epoch": 0.9910873440285205, "grad_norm": 2.71875, "learning_rate": 6.124346125233604e-08, "loss": 0.7707, "step": 8340 }, { "epoch": 0.9934640522875817, "grad_norm": 1.796875, "learning_rate": 3.2936926611149e-08, "loss": 0.6724, "step": 8360 }, { "epoch": 0.995840760546643, "grad_norm": 1.5390625, "learning_rate": 1.3338529967010793e-08, "loss": 0.6268, "step": 8380 }, { "epoch": 0.9982174688057041, "grad_norm": 3.828125, "learning_rate": 2.449978537655273e-09, "loss": 0.8173, "step": 8400 }, { "epoch": 1.0, "step": 8415, "total_flos": 9.688250026819584e+16, "train_loss": 1.1022938394574842, "train_runtime": 7319.3616, "train_samples_per_second": 1.15, "train_steps_per_second": 1.15 } ], "logging_steps": 20, "max_steps": 8415, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 9.688250026819584e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }