{ "best_global_step": 70000, "best_metric": 1.029943823814392, "best_model_checkpoint": "./dna_model/checkpoint-70000", "epoch": 10.0, "eval_steps": 5000, "global_step": 72380, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00013815971262779773, "grad_norm": 42.580928802490234, "learning_rate": 0.0, "loss": 1.6625, "step": 1 }, { "epoch": 0.013815971262779773, "grad_norm": 2.4457767009735107, "learning_rate": 2.97e-05, "loss": 1.36, "step": 100 }, { "epoch": 0.027631942525559547, "grad_norm": 0.5432274341583252, "learning_rate": 5.97e-05, "loss": 1.3309, "step": 200 }, { "epoch": 0.04144791378833932, "grad_norm": 0.825528621673584, "learning_rate": 8.969999999999998e-05, "loss": 1.3234, "step": 300 }, { "epoch": 0.055263885051119094, "grad_norm": 0.4912604093551636, "learning_rate": 0.0001197, "loss": 1.3249, "step": 400 }, { "epoch": 0.06907985631389886, "grad_norm": 0.9077563881874084, "learning_rate": 0.00014969999999999998, "loss": 1.3153, "step": 500 }, { "epoch": 0.08289582757667864, "grad_norm": 0.8954246640205383, "learning_rate": 0.00017969999999999998, "loss": 1.3123, "step": 600 }, { "epoch": 0.09671179883945841, "grad_norm": 0.5876831412315369, "learning_rate": 0.00020969999999999997, "loss": 1.3098, "step": 700 }, { "epoch": 0.11052777010223819, "grad_norm": 0.426789253950119, "learning_rate": 0.0002397, "loss": 1.3072, "step": 800 }, { "epoch": 0.12434374136501795, "grad_norm": 0.3324718177318573, "learning_rate": 0.0002697, "loss": 1.3037, "step": 900 }, { "epoch": 0.13815971262779772, "grad_norm": 0.23672613501548767, "learning_rate": 0.00029969999999999997, "loss": 1.2991, "step": 1000 }, { "epoch": 0.1519756838905775, "grad_norm": 0.4699796438217163, "learning_rate": 0.00029958391706360325, "loss": 1.2923, "step": 1100 }, { "epoch": 0.16579165515335728, "grad_norm": 0.684186577796936, "learning_rate": 0.00029916363126926307, "loss": 1.2825, "step": 1200 }, { "epoch": 0.17960762641613706, "grad_norm": 0.3944641649723053, "learning_rate": 0.00029874334547492294, "loss": 1.2678, "step": 1300 }, { "epoch": 0.19342359767891681, "grad_norm": 1.1556001901626587, "learning_rate": 0.00029832305968058276, "loss": 1.2541, "step": 1400 }, { "epoch": 0.2072395689416966, "grad_norm": 0.39745599031448364, "learning_rate": 0.0002979027738862426, "loss": 1.2439, "step": 1500 }, { "epoch": 0.22105554020447638, "grad_norm": 0.5201444029808044, "learning_rate": 0.00029748248809190246, "loss": 1.2329, "step": 1600 }, { "epoch": 0.23487151146725616, "grad_norm": 0.2168777734041214, "learning_rate": 0.00029706220229756234, "loss": 1.2268, "step": 1700 }, { "epoch": 0.2486874827300359, "grad_norm": 0.30599427223205566, "learning_rate": 0.00029664191650322216, "loss": 1.2199, "step": 1800 }, { "epoch": 0.2625034539928157, "grad_norm": 0.32062044739723206, "learning_rate": 0.00029622163070888203, "loss": 1.2131, "step": 1900 }, { "epoch": 0.27631942525559544, "grad_norm": 0.13411013782024384, "learning_rate": 0.00029580134491454186, "loss": 1.2074, "step": 2000 }, { "epoch": 0.2901353965183752, "grad_norm": 0.3672633767127991, "learning_rate": 0.00029538105912020173, "loss": 1.2022, "step": 2100 }, { "epoch": 0.303951367781155, "grad_norm": 0.41515815258026123, "learning_rate": 0.00029496077332586155, "loss": 1.1949, "step": 2200 }, { "epoch": 0.3177673390439348, "grad_norm": 0.18381068110466003, "learning_rate": 0.0002945404875315214, "loss": 1.1887, "step": 2300 }, { "epoch": 0.33158331030671456, "grad_norm": 0.3080751895904541, "learning_rate": 0.00029412020173718125, "loss": 1.1844, "step": 2400 }, { "epoch": 0.34539928156949434, "grad_norm": 0.38037416338920593, "learning_rate": 0.0002936999159428411, "loss": 1.1804, "step": 2500 }, { "epoch": 0.3592152528322741, "grad_norm": 0.23272989690303802, "learning_rate": 0.00029327963014850095, "loss": 1.1753, "step": 2600 }, { "epoch": 0.3730312240950539, "grad_norm": 0.1149936243891716, "learning_rate": 0.0002928593443541608, "loss": 1.1739, "step": 2700 }, { "epoch": 0.38684719535783363, "grad_norm": 0.28469276428222656, "learning_rate": 0.00029243905855982064, "loss": 1.1671, "step": 2800 }, { "epoch": 0.4006631666206134, "grad_norm": 0.25204166769981384, "learning_rate": 0.0002920187727654805, "loss": 1.1633, "step": 2900 }, { "epoch": 0.4144791378833932, "grad_norm": 0.3945861756801605, "learning_rate": 0.00029159848697114034, "loss": 1.1608, "step": 3000 }, { "epoch": 0.42829510914617297, "grad_norm": 0.2578865587711334, "learning_rate": 0.00029117820117680016, "loss": 1.1622, "step": 3100 }, { "epoch": 0.44211108040895275, "grad_norm": 0.16060177981853485, "learning_rate": 0.00029075791538246004, "loss": 1.1577, "step": 3200 }, { "epoch": 0.45592705167173253, "grad_norm": 0.1980718970298767, "learning_rate": 0.0002903376295881199, "loss": 1.155, "step": 3300 }, { "epoch": 0.4697430229345123, "grad_norm": 0.12515653669834137, "learning_rate": 0.00028991734379377974, "loss": 1.1519, "step": 3400 }, { "epoch": 0.4835589941972921, "grad_norm": 0.26255738735198975, "learning_rate": 0.0002894970579994396, "loss": 1.1523, "step": 3500 }, { "epoch": 0.4973749654600718, "grad_norm": 0.281464546918869, "learning_rate": 0.00028907677220509943, "loss": 1.1511, "step": 3600 }, { "epoch": 0.5111909367228517, "grad_norm": 0.11816036701202393, "learning_rate": 0.0002886564864107593, "loss": 1.1469, "step": 3700 }, { "epoch": 0.5250069079856314, "grad_norm": 0.25923675298690796, "learning_rate": 0.00028823620061641913, "loss": 1.1456, "step": 3800 }, { "epoch": 0.5388228792484112, "grad_norm": 0.2766472399234772, "learning_rate": 0.00028781591482207895, "loss": 1.1442, "step": 3900 }, { "epoch": 0.5526388505111909, "grad_norm": 0.1701624095439911, "learning_rate": 0.00028739562902773883, "loss": 1.1445, "step": 4000 }, { "epoch": 0.5664548217739707, "grad_norm": 0.3141656219959259, "learning_rate": 0.0002869753432333987, "loss": 1.1392, "step": 4100 }, { "epoch": 0.5802707930367504, "grad_norm": 0.11816743016242981, "learning_rate": 0.0002865550574390585, "loss": 1.1406, "step": 4200 }, { "epoch": 0.5940867642995302, "grad_norm": 0.12762723863124847, "learning_rate": 0.0002861347716447184, "loss": 1.1361, "step": 4300 }, { "epoch": 0.60790273556231, "grad_norm": 0.09322622418403625, "learning_rate": 0.0002857144858503782, "loss": 1.134, "step": 4400 }, { "epoch": 0.6217187068250898, "grad_norm": 0.1586735099554062, "learning_rate": 0.0002852942000560381, "loss": 1.1336, "step": 4500 }, { "epoch": 0.6355346780878696, "grad_norm": 0.13594642281532288, "learning_rate": 0.0002848739142616979, "loss": 1.1328, "step": 4600 }, { "epoch": 0.6493506493506493, "grad_norm": 0.21865279972553253, "learning_rate": 0.00028445362846735774, "loss": 1.1311, "step": 4700 }, { "epoch": 0.6631666206134291, "grad_norm": 0.22787001729011536, "learning_rate": 0.0002840333426730176, "loss": 1.1271, "step": 4800 }, { "epoch": 0.6769825918762089, "grad_norm": 0.2334531843662262, "learning_rate": 0.0002836130568786775, "loss": 1.1291, "step": 4900 }, { "epoch": 0.6907985631389887, "grad_norm": 0.11103236675262451, "learning_rate": 0.0002831927710843373, "loss": 1.1252, "step": 5000 }, { "epoch": 0.6907985631389887, "eval_accuracy": 0.4745045939970608, "eval_loss": 1.1205766201019287, "eval_runtime": 1027.9902, "eval_samples_per_second": 200.256, "eval_steps_per_second": 6.259, "step": 5000 }, { "epoch": 0.7046145344017685, "grad_norm": 0.21742330491542816, "learning_rate": 0.0002827724852899972, "loss": 1.1235, "step": 5100 }, { "epoch": 0.7184305056645482, "grad_norm": 0.23728515207767487, "learning_rate": 0.000282352199495657, "loss": 1.1233, "step": 5200 }, { "epoch": 0.732246476927328, "grad_norm": 0.21022765338420868, "learning_rate": 0.0002819319137013169, "loss": 1.1236, "step": 5300 }, { "epoch": 0.7460624481901078, "grad_norm": 0.0924484059214592, "learning_rate": 0.0002815116279069767, "loss": 1.1215, "step": 5400 }, { "epoch": 0.7598784194528876, "grad_norm": 0.1716778427362442, "learning_rate": 0.00028109134211263653, "loss": 1.1238, "step": 5500 }, { "epoch": 0.7736943907156673, "grad_norm": 0.13049638271331787, "learning_rate": 0.0002806710563182964, "loss": 1.1185, "step": 5600 }, { "epoch": 0.787510361978447, "grad_norm": 0.16255174577236176, "learning_rate": 0.0002802507705239563, "loss": 1.1169, "step": 5700 }, { "epoch": 0.8013263332412268, "grad_norm": 0.10065080225467682, "learning_rate": 0.0002798304847296161, "loss": 1.1184, "step": 5800 }, { "epoch": 0.8151423045040066, "grad_norm": 0.1182553768157959, "learning_rate": 0.000279410198935276, "loss": 1.1141, "step": 5900 }, { "epoch": 0.8289582757667864, "grad_norm": 0.14556263387203217, "learning_rate": 0.0002789899131409358, "loss": 1.1154, "step": 6000 }, { "epoch": 0.8427742470295662, "grad_norm": 0.1383764147758484, "learning_rate": 0.00027857383020453907, "loss": 1.1118, "step": 6100 }, { "epoch": 0.8565902182923459, "grad_norm": 0.2821154296398163, "learning_rate": 0.00027815354441019895, "loss": 1.1104, "step": 6200 }, { "epoch": 0.8704061895551257, "grad_norm": 0.22286450862884521, "learning_rate": 0.00027773325861585877, "loss": 1.1109, "step": 6300 }, { "epoch": 0.8842221608179055, "grad_norm": 0.2058987319469452, "learning_rate": 0.0002773129728215186, "loss": 1.1093, "step": 6400 }, { "epoch": 0.8980381320806853, "grad_norm": 0.21338045597076416, "learning_rate": 0.00027689268702717847, "loss": 1.1091, "step": 6500 }, { "epoch": 0.9118541033434651, "grad_norm": 0.0900028795003891, "learning_rate": 0.0002764724012328383, "loss": 1.1067, "step": 6600 }, { "epoch": 0.9256700746062448, "grad_norm": 0.10679551959037781, "learning_rate": 0.00027605211543849816, "loss": 1.108, "step": 6700 }, { "epoch": 0.9394860458690246, "grad_norm": 0.07972779124975204, "learning_rate": 0.000275631829644158, "loss": 1.1057, "step": 6800 }, { "epoch": 0.9533020171318044, "grad_norm": 0.24500218033790588, "learning_rate": 0.00027521154384981786, "loss": 1.105, "step": 6900 }, { "epoch": 0.9671179883945842, "grad_norm": 0.11576998978853226, "learning_rate": 0.00027479125805547774, "loss": 1.1029, "step": 7000 }, { "epoch": 0.980933959657364, "grad_norm": 0.10553757101297379, "learning_rate": 0.00027437097226113756, "loss": 1.1041, "step": 7100 }, { "epoch": 0.9947499309201436, "grad_norm": 0.15332186222076416, "learning_rate": 0.0002739506864667974, "loss": 1.0982, "step": 7200 }, { "epoch": 1.0085659021829234, "grad_norm": 0.11897014081478119, "learning_rate": 0.00027353040067245725, "loss": 1.0996, "step": 7300 }, { "epoch": 1.0223818734457033, "grad_norm": 0.1156444102525711, "learning_rate": 0.0002731101148781171, "loss": 1.1032, "step": 7400 }, { "epoch": 1.036197844708483, "grad_norm": 0.06223931908607483, "learning_rate": 0.00027268982908377695, "loss": 1.0982, "step": 7500 }, { "epoch": 1.0500138159712629, "grad_norm": 0.14377152919769287, "learning_rate": 0.00027226954328943677, "loss": 1.1003, "step": 7600 }, { "epoch": 1.0638297872340425, "grad_norm": 0.12667153775691986, "learning_rate": 0.00027184925749509665, "loss": 1.0989, "step": 7700 }, { "epoch": 1.0776457584968224, "grad_norm": 0.16101804375648499, "learning_rate": 0.0002714289717007565, "loss": 1.0968, "step": 7800 }, { "epoch": 1.091461729759602, "grad_norm": 0.06424383819103241, "learning_rate": 0.00027100868590641635, "loss": 1.0955, "step": 7900 }, { "epoch": 1.105277701022382, "grad_norm": 0.09638939052820206, "learning_rate": 0.00027058840011207617, "loss": 1.095, "step": 8000 }, { "epoch": 1.1190936722851617, "grad_norm": 0.08098015189170837, "learning_rate": 0.00027016811431773604, "loss": 1.0969, "step": 8100 }, { "epoch": 1.1329096435479413, "grad_norm": 0.10837887227535248, "learning_rate": 0.00026974782852339586, "loss": 1.096, "step": 8200 }, { "epoch": 1.1467256148107212, "grad_norm": 0.05644046515226364, "learning_rate": 0.00026932754272905574, "loss": 1.0944, "step": 8300 }, { "epoch": 1.1605415860735009, "grad_norm": 0.12965446710586548, "learning_rate": 0.00026890725693471556, "loss": 1.0953, "step": 8400 }, { "epoch": 1.1743575573362808, "grad_norm": 0.12333771586418152, "learning_rate": 0.00026848697114037544, "loss": 1.095, "step": 8500 }, { "epoch": 1.1881735285990604, "grad_norm": 0.1270703673362732, "learning_rate": 0.0002680666853460353, "loss": 1.0929, "step": 8600 }, { "epoch": 1.2019894998618403, "grad_norm": 0.16918766498565674, "learning_rate": 0.00026764639955169513, "loss": 1.0918, "step": 8700 }, { "epoch": 1.21580547112462, "grad_norm": 0.08776108920574188, "learning_rate": 0.00026722611375735496, "loss": 1.0952, "step": 8800 }, { "epoch": 1.2296214423874, "grad_norm": 0.08252176642417908, "learning_rate": 0.00026680582796301483, "loss": 1.09, "step": 8900 }, { "epoch": 1.2434374136501796, "grad_norm": 0.16331979632377625, "learning_rate": 0.00026638554216867465, "loss": 1.0898, "step": 9000 }, { "epoch": 1.2572533849129595, "grad_norm": 0.17065368592739105, "learning_rate": 0.00026596525637433453, "loss": 1.0907, "step": 9100 }, { "epoch": 1.2710693561757391, "grad_norm": 0.12038784474134445, "learning_rate": 0.00026554497057999435, "loss": 1.0856, "step": 9200 }, { "epoch": 1.284885327438519, "grad_norm": 0.11924347281455994, "learning_rate": 0.0002651246847856542, "loss": 1.0895, "step": 9300 }, { "epoch": 1.2987012987012987, "grad_norm": 0.1443828046321869, "learning_rate": 0.0002647043989913141, "loss": 1.0874, "step": 9400 }, { "epoch": 1.3125172699640784, "grad_norm": 0.14472317695617676, "learning_rate": 0.0002642841131969739, "loss": 1.0879, "step": 9500 }, { "epoch": 1.3263332412268583, "grad_norm": 0.15847088396549225, "learning_rate": 0.00026386382740263374, "loss": 1.0873, "step": 9600 }, { "epoch": 1.3401492124896381, "grad_norm": 0.17960332334041595, "learning_rate": 0.0002634435416082936, "loss": 1.0887, "step": 9700 }, { "epoch": 1.3539651837524178, "grad_norm": 0.1566227227449417, "learning_rate": 0.00026302325581395344, "loss": 1.0884, "step": 9800 }, { "epoch": 1.3677811550151975, "grad_norm": 0.1431213617324829, "learning_rate": 0.0002626029700196133, "loss": 1.0864, "step": 9900 }, { "epoch": 1.3815971262779774, "grad_norm": 0.10321222990751266, "learning_rate": 0.0002621826842252732, "loss": 1.0835, "step": 10000 }, { "epoch": 1.3815971262779774, "eval_accuracy": 0.49913821881815945, "eval_loss": 1.081355094909668, "eval_runtime": 748.8314, "eval_samples_per_second": 274.91, "eval_steps_per_second": 8.592, "step": 10000 }, { "epoch": 1.395413097540757, "grad_norm": 0.10260605067014694, "learning_rate": 0.0002617666012888764, "loss": 1.0843, "step": 10100 }, { "epoch": 1.409229068803537, "grad_norm": 0.1076885387301445, "learning_rate": 0.0002613463154945363, "loss": 1.0845, "step": 10200 }, { "epoch": 1.4230450400663166, "grad_norm": 0.0723571702837944, "learning_rate": 0.0002609260297001961, "loss": 1.0814, "step": 10300 }, { "epoch": 1.4368610113290965, "grad_norm": 0.10695687681436539, "learning_rate": 0.00026050574390585593, "loss": 1.0842, "step": 10400 }, { "epoch": 1.4506769825918762, "grad_norm": 0.11008185893297195, "learning_rate": 0.0002600854581115158, "loss": 1.0832, "step": 10500 }, { "epoch": 1.464492953854656, "grad_norm": 0.12239653617143631, "learning_rate": 0.0002596651723171756, "loss": 1.0813, "step": 10600 }, { "epoch": 1.4783089251174357, "grad_norm": 0.11045056581497192, "learning_rate": 0.0002592448865228355, "loss": 1.0848, "step": 10700 }, { "epoch": 1.4921248963802154, "grad_norm": 0.07234488427639008, "learning_rate": 0.0002588246007284954, "loss": 1.0826, "step": 10800 }, { "epoch": 1.5059408676429953, "grad_norm": 0.11086778342723846, "learning_rate": 0.0002584043149341552, "loss": 1.0804, "step": 10900 }, { "epoch": 1.5197568389057752, "grad_norm": 0.10693442821502686, "learning_rate": 0.0002579840291398151, "loss": 1.0784, "step": 11000 }, { "epoch": 1.5335728101685548, "grad_norm": 0.11604110896587372, "learning_rate": 0.0002575637433454749, "loss": 1.0792, "step": 11100 }, { "epoch": 1.5473887814313345, "grad_norm": 0.0809662714600563, "learning_rate": 0.0002571434575511347, "loss": 1.083, "step": 11200 }, { "epoch": 1.5612047526941144, "grad_norm": 0.1850002408027649, "learning_rate": 0.0002567231717567946, "loss": 1.0802, "step": 11300 }, { "epoch": 1.5750207239568943, "grad_norm": 0.0779227465391159, "learning_rate": 0.0002563028859624544, "loss": 1.0811, "step": 11400 }, { "epoch": 1.588836695219674, "grad_norm": 0.16764625906944275, "learning_rate": 0.0002558826001681143, "loss": 1.0763, "step": 11500 }, { "epoch": 1.6026526664824536, "grad_norm": 0.11104313284158707, "learning_rate": 0.00025546231437377417, "loss": 1.0782, "step": 11600 }, { "epoch": 1.6164686377452335, "grad_norm": 0.16667212545871735, "learning_rate": 0.000255042028579434, "loss": 1.0781, "step": 11700 }, { "epoch": 1.6302846090080134, "grad_norm": 0.2246047705411911, "learning_rate": 0.00025462174278509386, "loss": 1.08, "step": 11800 }, { "epoch": 1.644100580270793, "grad_norm": 0.2305343896150589, "learning_rate": 0.0002542014569907537, "loss": 1.0756, "step": 11900 }, { "epoch": 1.6579165515335728, "grad_norm": 0.13618823885917664, "learning_rate": 0.0002537811711964135, "loss": 1.076, "step": 12000 }, { "epoch": 1.6717325227963524, "grad_norm": 0.15795475244522095, "learning_rate": 0.0002533608854020734, "loss": 1.0749, "step": 12100 }, { "epoch": 1.6855484940591323, "grad_norm": 0.20267115533351898, "learning_rate": 0.00025294480246567665, "loss": 1.077, "step": 12200 }, { "epoch": 1.6993644653219122, "grad_norm": 0.08052489906549454, "learning_rate": 0.0002525245166713365, "loss": 1.073, "step": 12300 }, { "epoch": 1.7131804365846919, "grad_norm": 0.11914093047380447, "learning_rate": 0.00025210423087699635, "loss": 1.0755, "step": 12400 }, { "epoch": 1.7269964078474715, "grad_norm": 0.12703542411327362, "learning_rate": 0.00025168394508265617, "loss": 1.0765, "step": 12500 }, { "epoch": 1.7408123791102514, "grad_norm": 0.12948518991470337, "learning_rate": 0.00025126365928831605, "loss": 1.0748, "step": 12600 }, { "epoch": 1.7546283503730313, "grad_norm": 0.1027710810303688, "learning_rate": 0.00025084337349397587, "loss": 1.0745, "step": 12700 }, { "epoch": 1.768444321635811, "grad_norm": 0.20131652057170868, "learning_rate": 0.0002504230876996357, "loss": 1.0731, "step": 12800 }, { "epoch": 1.7822602928985907, "grad_norm": 0.0673370212316513, "learning_rate": 0.00025000280190529557, "loss": 1.0721, "step": 12900 }, { "epoch": 1.7960762641613706, "grad_norm": 0.10322799533605576, "learning_rate": 0.00024958251611095544, "loss": 1.0731, "step": 13000 }, { "epoch": 1.8098922354241505, "grad_norm": 0.08498311042785645, "learning_rate": 0.00024916223031661526, "loss": 1.0722, "step": 13100 }, { "epoch": 1.8237082066869301, "grad_norm": 0.07025079429149628, "learning_rate": 0.00024874194452227514, "loss": 1.0725, "step": 13200 }, { "epoch": 1.8375241779497098, "grad_norm": 0.13933932781219482, "learning_rate": 0.00024832165872793496, "loss": 1.0714, "step": 13300 }, { "epoch": 1.8513401492124897, "grad_norm": 0.10513993352651596, "learning_rate": 0.00024790137293359484, "loss": 1.0725, "step": 13400 }, { "epoch": 1.8651561204752696, "grad_norm": 0.1704607903957367, "learning_rate": 0.0002474810871392547, "loss": 1.0712, "step": 13500 }, { "epoch": 1.8789720917380492, "grad_norm": 0.08315689861774445, "learning_rate": 0.0002470608013449145, "loss": 1.0697, "step": 13600 }, { "epoch": 1.892788063000829, "grad_norm": 0.09900273382663727, "learning_rate": 0.00024664051555057436, "loss": 1.0735, "step": 13700 }, { "epoch": 1.9066040342636086, "grad_norm": 0.05560864508152008, "learning_rate": 0.00024622022975623423, "loss": 1.0711, "step": 13800 }, { "epoch": 1.9204200055263885, "grad_norm": 0.13863462209701538, "learning_rate": 0.00024579994396189405, "loss": 1.0681, "step": 13900 }, { "epoch": 1.9342359767891684, "grad_norm": 0.07841744273900986, "learning_rate": 0.00024537965816755393, "loss": 1.0711, "step": 14000 }, { "epoch": 1.948051948051948, "grad_norm": 0.058312736451625824, "learning_rate": 0.00024495937237321375, "loss": 1.0709, "step": 14100 }, { "epoch": 1.9618679193147277, "grad_norm": 0.11208023875951767, "learning_rate": 0.000244543289436817, "loss": 1.0686, "step": 14200 }, { "epoch": 1.9756838905775076, "grad_norm": 0.10133163630962372, "learning_rate": 0.00024412300364247687, "loss": 1.0683, "step": 14300 }, { "epoch": 1.9894998618402875, "grad_norm": 0.08370282500982285, "learning_rate": 0.0002437027178481367, "loss": 1.0709, "step": 14400 }, { "epoch": 2.003315833103067, "grad_norm": 0.09476770460605621, "learning_rate": 0.00024328243205379654, "loss": 1.0697, "step": 14500 }, { "epoch": 2.017131804365847, "grad_norm": 0.0733637660741806, "learning_rate": 0.0002428621462594564, "loss": 1.0681, "step": 14600 }, { "epoch": 2.0309477756286265, "grad_norm": 0.09925834089517593, "learning_rate": 0.00024244186046511627, "loss": 1.0702, "step": 14700 }, { "epoch": 2.0447637468914066, "grad_norm": 0.15911750495433807, "learning_rate": 0.00024202157467077611, "loss": 1.0665, "step": 14800 }, { "epoch": 2.0585797181541863, "grad_norm": 0.13638247549533844, "learning_rate": 0.00024160128887643596, "loss": 1.0696, "step": 14900 }, { "epoch": 2.072395689416966, "grad_norm": 0.16883982717990875, "learning_rate": 0.0002411810030820958, "loss": 1.0641, "step": 15000 }, { "epoch": 2.072395689416966, "eval_accuracy": 0.5102966510685876, "eval_loss": 1.0638896226882935, "eval_runtime": 924.2494, "eval_samples_per_second": 222.733, "eval_steps_per_second": 6.961, "step": 15000 }, { "epoch": 2.0862116606797456, "grad_norm": 0.09925784170627594, "learning_rate": 0.00024076071728775566, "loss": 1.0683, "step": 15100 }, { "epoch": 2.1000276319425257, "grad_norm": 0.06180203706026077, "learning_rate": 0.00024034043149341548, "loss": 1.066, "step": 15200 }, { "epoch": 2.1138436032053054, "grad_norm": 0.10063247382640839, "learning_rate": 0.00023992014569907533, "loss": 1.0668, "step": 15300 }, { "epoch": 2.127659574468085, "grad_norm": 0.11476041376590729, "learning_rate": 0.0002394998599047352, "loss": 1.0644, "step": 15400 }, { "epoch": 2.1414755457308647, "grad_norm": 0.11798429489135742, "learning_rate": 0.00023907957411039505, "loss": 1.0626, "step": 15500 }, { "epoch": 2.155291516993645, "grad_norm": 0.13165287673473358, "learning_rate": 0.0002386592883160549, "loss": 1.0648, "step": 15600 }, { "epoch": 2.1691074882564245, "grad_norm": 0.1705123484134674, "learning_rate": 0.00023823900252171475, "loss": 1.0639, "step": 15700 }, { "epoch": 2.182923459519204, "grad_norm": 0.13375049829483032, "learning_rate": 0.0002378187167273746, "loss": 1.062, "step": 15800 }, { "epoch": 2.196739430781984, "grad_norm": 0.09405038505792618, "learning_rate": 0.00023739843093303445, "loss": 1.0634, "step": 15900 }, { "epoch": 2.210555402044764, "grad_norm": 0.11285752803087234, "learning_rate": 0.00023697814513869427, "loss": 1.0667, "step": 16000 }, { "epoch": 2.2243713733075436, "grad_norm": 0.12377699464559555, "learning_rate": 0.00023655785934435412, "loss": 1.064, "step": 16100 }, { "epoch": 2.2381873445703233, "grad_norm": 0.0979316234588623, "learning_rate": 0.000236137573550014, "loss": 1.0621, "step": 16200 }, { "epoch": 2.252003315833103, "grad_norm": 0.11494515091180801, "learning_rate": 0.00023572149061361724, "loss": 1.0645, "step": 16300 }, { "epoch": 2.2658192870958827, "grad_norm": 0.07066236436367035, "learning_rate": 0.0002353012048192771, "loss": 1.063, "step": 16400 }, { "epoch": 2.2796352583586628, "grad_norm": 0.08686563372612, "learning_rate": 0.00023488091902493694, "loss": 1.066, "step": 16500 }, { "epoch": 2.2934512296214424, "grad_norm": 0.058148209005594254, "learning_rate": 0.00023446063323059678, "loss": 1.0643, "step": 16600 }, { "epoch": 2.307267200884222, "grad_norm": 0.14033359289169312, "learning_rate": 0.00023404034743625666, "loss": 1.0634, "step": 16700 }, { "epoch": 2.3210831721470018, "grad_norm": 0.09940097481012344, "learning_rate": 0.00023362006164191645, "loss": 1.0629, "step": 16800 }, { "epoch": 2.334899143409782, "grad_norm": 0.08228994905948639, "learning_rate": 0.00023319977584757633, "loss": 1.0626, "step": 16900 }, { "epoch": 2.3487151146725616, "grad_norm": 0.05418753623962402, "learning_rate": 0.00023277949005323618, "loss": 1.0611, "step": 17000 }, { "epoch": 2.3625310859353412, "grad_norm": 0.09691222757101059, "learning_rate": 0.00023235920425889603, "loss": 1.0626, "step": 17100 }, { "epoch": 2.376347057198121, "grad_norm": 0.1607312560081482, "learning_rate": 0.00023193891846455588, "loss": 1.0623, "step": 17200 }, { "epoch": 2.3901630284609006, "grad_norm": 0.1193649098277092, "learning_rate": 0.00023151863267021572, "loss": 1.0627, "step": 17300 }, { "epoch": 2.4039789997236807, "grad_norm": 0.05427398905158043, "learning_rate": 0.00023109834687587557, "loss": 1.0609, "step": 17400 }, { "epoch": 2.4177949709864603, "grad_norm": 0.10591702163219452, "learning_rate": 0.00023067806108153545, "loss": 1.0637, "step": 17500 }, { "epoch": 2.43161094224924, "grad_norm": 0.057032886892557144, "learning_rate": 0.00023025777528719524, "loss": 1.0612, "step": 17600 }, { "epoch": 2.44542691351202, "grad_norm": 0.08455175161361694, "learning_rate": 0.00022983748949285512, "loss": 1.0606, "step": 17700 }, { "epoch": 2.4592428847748, "grad_norm": 0.13975144922733307, "learning_rate": 0.00022941720369851497, "loss": 1.0624, "step": 17800 }, { "epoch": 2.4730588560375795, "grad_norm": 0.11535393446683884, "learning_rate": 0.00022899691790417482, "loss": 1.0603, "step": 17900 }, { "epoch": 2.486874827300359, "grad_norm": 0.10047648102045059, "learning_rate": 0.00022857663210983466, "loss": 1.0607, "step": 18000 }, { "epoch": 2.500690798563139, "grad_norm": 0.08474704623222351, "learning_rate": 0.0002281563463154945, "loss": 1.062, "step": 18100 }, { "epoch": 2.514506769825919, "grad_norm": 0.15308576822280884, "learning_rate": 0.00022773606052115436, "loss": 1.0603, "step": 18200 }, { "epoch": 2.5283227410886986, "grad_norm": 0.05684039369225502, "learning_rate": 0.00022731577472681424, "loss": 1.0589, "step": 18300 }, { "epoch": 2.5421387123514783, "grad_norm": 0.10712555050849915, "learning_rate": 0.00022689548893247409, "loss": 1.0592, "step": 18400 }, { "epoch": 2.555954683614258, "grad_norm": 0.0800655260682106, "learning_rate": 0.0002264794059960773, "loss": 1.0603, "step": 18500 }, { "epoch": 2.569770654877038, "grad_norm": 0.05980188027024269, "learning_rate": 0.00022605912020173715, "loss": 1.0608, "step": 18600 }, { "epoch": 2.5835866261398177, "grad_norm": 0.052051473408937454, "learning_rate": 0.000225638834407397, "loss": 1.0603, "step": 18700 }, { "epoch": 2.5974025974025974, "grad_norm": 0.11966883391141891, "learning_rate": 0.00022521854861305685, "loss": 1.057, "step": 18800 }, { "epoch": 2.611218568665377, "grad_norm": 0.08861220628023148, "learning_rate": 0.00022479826281871673, "loss": 1.0603, "step": 18900 }, { "epoch": 2.6250345399281567, "grad_norm": 0.12264814227819443, "learning_rate": 0.00022437797702437657, "loss": 1.0602, "step": 19000 }, { "epoch": 2.638850511190937, "grad_norm": 0.08384163677692413, "learning_rate": 0.00022395769123003642, "loss": 1.057, "step": 19100 }, { "epoch": 2.6526664824537165, "grad_norm": 0.11168386787176132, "learning_rate": 0.00022353740543569624, "loss": 1.0572, "step": 19200 }, { "epoch": 2.666482453716496, "grad_norm": 0.12558519840240479, "learning_rate": 0.0002231171196413561, "loss": 1.0592, "step": 19300 }, { "epoch": 2.6802984249792763, "grad_norm": 0.06810207664966583, "learning_rate": 0.00022269683384701594, "loss": 1.055, "step": 19400 }, { "epoch": 2.694114396242056, "grad_norm": 0.16571113467216492, "learning_rate": 0.0002222765480526758, "loss": 1.0599, "step": 19500 }, { "epoch": 2.7079303675048356, "grad_norm": 0.07613151520490646, "learning_rate": 0.00022185626225833564, "loss": 1.0564, "step": 19600 }, { "epoch": 2.7217463387676153, "grad_norm": 0.08713393658399582, "learning_rate": 0.00022143597646399551, "loss": 1.0582, "step": 19700 }, { "epoch": 2.735562310030395, "grad_norm": 0.11707925796508789, "learning_rate": 0.00022101569066965536, "loss": 1.056, "step": 19800 }, { "epoch": 2.749378281293175, "grad_norm": 0.1053171455860138, "learning_rate": 0.0002205954048753152, "loss": 1.0608, "step": 19900 }, { "epoch": 2.7631942525559547, "grad_norm": 0.056531500071287155, "learning_rate": 0.00022017511908097506, "loss": 1.0563, "step": 20000 }, { "epoch": 2.7631942525559547, "eval_accuracy": 0.516310033016185, "eval_loss": 1.054749608039856, "eval_runtime": 731.5154, "eval_samples_per_second": 281.417, "eval_steps_per_second": 8.795, "step": 20000 }, { "epoch": 2.7770102238187344, "grad_norm": 0.10811367630958557, "learning_rate": 0.00021975483328663488, "loss": 1.0556, "step": 20100 }, { "epoch": 2.790826195081514, "grad_norm": 0.06601472198963165, "learning_rate": 0.00021933454749229473, "loss": 1.0578, "step": 20200 }, { "epoch": 2.804642166344294, "grad_norm": 0.06906837224960327, "learning_rate": 0.00021891426169795458, "loss": 1.06, "step": 20300 }, { "epoch": 2.818458137607074, "grad_norm": 0.08911406248807907, "learning_rate": 0.00021849397590361443, "loss": 1.0583, "step": 20400 }, { "epoch": 2.8322741088698535, "grad_norm": 0.06497912108898163, "learning_rate": 0.0002180778929672177, "loss": 1.0575, "step": 20500 }, { "epoch": 2.846090080132633, "grad_norm": 0.0886107012629509, "learning_rate": 0.00021765760717287755, "loss": 1.0552, "step": 20600 }, { "epoch": 2.859906051395413, "grad_norm": 0.05942055955529213, "learning_rate": 0.0002172373213785374, "loss": 1.0533, "step": 20700 }, { "epoch": 2.873722022658193, "grad_norm": 0.13015809655189514, "learning_rate": 0.00021681703558419725, "loss": 1.0549, "step": 20800 }, { "epoch": 2.8875379939209727, "grad_norm": 0.06085093691945076, "learning_rate": 0.00021639674978985707, "loss": 1.057, "step": 20900 }, { "epoch": 2.9013539651837523, "grad_norm": 0.17039401829242706, "learning_rate": 0.00021597646399551692, "loss": 1.0571, "step": 21000 }, { "epoch": 2.9151699364465324, "grad_norm": 0.07950026541948318, "learning_rate": 0.00021555617820117676, "loss": 1.0535, "step": 21100 }, { "epoch": 2.928985907709312, "grad_norm": 0.1195695698261261, "learning_rate": 0.00021513589240683664, "loss": 1.0535, "step": 21200 }, { "epoch": 2.942801878972092, "grad_norm": 0.0896124541759491, "learning_rate": 0.0002147156066124965, "loss": 1.0534, "step": 21300 }, { "epoch": 2.9566178502348714, "grad_norm": 0.07629978656768799, "learning_rate": 0.00021429532081815634, "loss": 1.0564, "step": 21400 }, { "epoch": 2.970433821497651, "grad_norm": 0.07431907206773758, "learning_rate": 0.00021387503502381618, "loss": 1.0559, "step": 21500 }, { "epoch": 2.984249792760431, "grad_norm": 0.0771278440952301, "learning_rate": 0.00021345474922947603, "loss": 1.0562, "step": 21600 }, { "epoch": 2.998065764023211, "grad_norm": 0.11643990874290466, "learning_rate": 0.00021303446343513585, "loss": 1.0525, "step": 21700 }, { "epoch": 3.0118817352859906, "grad_norm": 0.058162059634923935, "learning_rate": 0.0002126141776407957, "loss": 1.0509, "step": 21800 }, { "epoch": 3.0256977065487702, "grad_norm": 0.12037301808595657, "learning_rate": 0.00021219389184645558, "loss": 1.0513, "step": 21900 }, { "epoch": 3.0395136778115504, "grad_norm": 0.052515506744384766, "learning_rate": 0.00021177360605211543, "loss": 1.051, "step": 22000 }, { "epoch": 3.05332964907433, "grad_norm": 0.10646827518939972, "learning_rate": 0.00021135332025777528, "loss": 1.0542, "step": 22100 }, { "epoch": 3.0671456203371097, "grad_norm": 0.1113181784749031, "learning_rate": 0.00021093303446343512, "loss": 1.0531, "step": 22200 }, { "epoch": 3.0809615915998894, "grad_norm": 0.07355222851037979, "learning_rate": 0.00021051274866909497, "loss": 1.0524, "step": 22300 }, { "epoch": 3.094777562862669, "grad_norm": 0.06925370544195175, "learning_rate": 0.00021009246287475482, "loss": 1.0535, "step": 22400 }, { "epoch": 3.108593534125449, "grad_norm": 0.048475924879312515, "learning_rate": 0.00020967217708041464, "loss": 1.0564, "step": 22500 }, { "epoch": 3.122409505388229, "grad_norm": 0.08578319102525711, "learning_rate": 0.0002092518912860745, "loss": 1.0519, "step": 22600 }, { "epoch": 3.1362254766510085, "grad_norm": 0.08585724979639053, "learning_rate": 0.00020883160549173437, "loss": 1.0525, "step": 22700 }, { "epoch": 3.150041447913788, "grad_norm": 0.06518802791833878, "learning_rate": 0.00020841131969739422, "loss": 1.0543, "step": 22800 }, { "epoch": 3.1638574191765683, "grad_norm": 0.046030618250370026, "learning_rate": 0.00020799103390305406, "loss": 1.0525, "step": 22900 }, { "epoch": 3.177673390439348, "grad_norm": 0.04972764104604721, "learning_rate": 0.0002075707481087139, "loss": 1.0512, "step": 23000 }, { "epoch": 3.1914893617021276, "grad_norm": 0.11977583914995193, "learning_rate": 0.00020715046231437376, "loss": 1.052, "step": 23100 }, { "epoch": 3.2053053329649073, "grad_norm": 0.08040472120046616, "learning_rate": 0.0002067301765200336, "loss": 1.0491, "step": 23200 }, { "epoch": 3.2191213042276874, "grad_norm": 0.10473213344812393, "learning_rate": 0.00020630989072569343, "loss": 1.0525, "step": 23300 }, { "epoch": 3.232937275490467, "grad_norm": 0.0790744498372078, "learning_rate": 0.00020588960493135328, "loss": 1.0508, "step": 23400 }, { "epoch": 3.2467532467532467, "grad_norm": 0.12807689607143402, "learning_rate": 0.00020547352199495655, "loss": 1.0485, "step": 23500 }, { "epoch": 3.2605692180160264, "grad_norm": 0.10298227518796921, "learning_rate": 0.0002050532362006164, "loss": 1.049, "step": 23600 }, { "epoch": 3.2743851892788065, "grad_norm": 0.11504103243350983, "learning_rate": 0.00020463295040627625, "loss": 1.0511, "step": 23700 }, { "epoch": 3.288201160541586, "grad_norm": 0.05548229441046715, "learning_rate": 0.0002042126646119361, "loss": 1.0499, "step": 23800 }, { "epoch": 3.302017131804366, "grad_norm": 0.06242981553077698, "learning_rate": 0.00020379237881759595, "loss": 1.0543, "step": 23900 }, { "epoch": 3.3158331030671455, "grad_norm": 0.12101748585700989, "learning_rate": 0.00020337209302325582, "loss": 1.0482, "step": 24000 }, { "epoch": 3.329649074329925, "grad_norm": 0.09176388382911682, "learning_rate": 0.00020295180722891562, "loss": 1.0514, "step": 24100 }, { "epoch": 3.3434650455927053, "grad_norm": 0.08758760988712311, "learning_rate": 0.0002025315214345755, "loss": 1.0505, "step": 24200 }, { "epoch": 3.357281016855485, "grad_norm": 0.06818066537380219, "learning_rate": 0.00020211123564023534, "loss": 1.0511, "step": 24300 }, { "epoch": 3.3710969881182646, "grad_norm": 0.10384306311607361, "learning_rate": 0.0002016909498458952, "loss": 1.0513, "step": 24400 }, { "epoch": 3.3849129593810443, "grad_norm": 0.12452493607997894, "learning_rate": 0.00020127066405155504, "loss": 1.0502, "step": 24500 }, { "epoch": 3.3987289306438244, "grad_norm": 0.07460072636604309, "learning_rate": 0.0002008503782572149, "loss": 1.0526, "step": 24600 }, { "epoch": 3.412544901906604, "grad_norm": 0.1017543151974678, "learning_rate": 0.00020043009246287474, "loss": 1.0501, "step": 24700 }, { "epoch": 3.4263608731693838, "grad_norm": 0.0900358185172081, "learning_rate": 0.0002000098066685346, "loss": 1.0512, "step": 24800 }, { "epoch": 3.4401768444321634, "grad_norm": 0.10934050381183624, "learning_rate": 0.00019958952087419443, "loss": 1.0495, "step": 24900 }, { "epoch": 3.4539928156949435, "grad_norm": 0.0656353011727333, "learning_rate": 0.00019916923507985428, "loss": 1.0504, "step": 25000 }, { "epoch": 3.4539928156949435, "eval_accuracy": 0.520419659075542, "eval_loss": 1.0485948324203491, "eval_runtime": 728.0613, "eval_samples_per_second": 282.752, "eval_steps_per_second": 8.837, "step": 25000 }, { "epoch": 3.467808786957723, "grad_norm": 0.07246037572622299, "learning_rate": 0.00019874894928551413, "loss": 1.0493, "step": 25100 }, { "epoch": 3.481624758220503, "grad_norm": 0.14033739268779755, "learning_rate": 0.00019832866349117398, "loss": 1.05, "step": 25200 }, { "epoch": 3.4954407294832825, "grad_norm": 0.05688853561878204, "learning_rate": 0.00019790837769683383, "loss": 1.0509, "step": 25300 }, { "epoch": 3.5092567007460627, "grad_norm": 0.053916674107313156, "learning_rate": 0.00019748809190249368, "loss": 1.0503, "step": 25400 }, { "epoch": 3.5230726720088423, "grad_norm": 0.12233688682317734, "learning_rate": 0.00019706780610815352, "loss": 1.05, "step": 25500 }, { "epoch": 3.536888643271622, "grad_norm": 0.10314755886793137, "learning_rate": 0.0001966475203138134, "loss": 1.0501, "step": 25600 }, { "epoch": 3.5507046145344017, "grad_norm": 0.05037887394428253, "learning_rate": 0.00019623143737741662, "loss": 1.0468, "step": 25700 }, { "epoch": 3.5645205857971813, "grad_norm": 0.13344399631023407, "learning_rate": 0.00019581115158307647, "loss": 1.0477, "step": 25800 }, { "epoch": 3.5783365570599615, "grad_norm": 0.07191654294729233, "learning_rate": 0.00019539086578873632, "loss": 1.0498, "step": 25900 }, { "epoch": 3.592152528322741, "grad_norm": 0.05592725798487663, "learning_rate": 0.00019497057999439616, "loss": 1.0506, "step": 26000 }, { "epoch": 3.605968499585521, "grad_norm": 0.10346696525812149, "learning_rate": 0.000194550294200056, "loss": 1.0499, "step": 26100 }, { "epoch": 3.619784470848301, "grad_norm": 0.09233855456113815, "learning_rate": 0.0001941300084057159, "loss": 1.0456, "step": 26200 }, { "epoch": 3.6336004421110806, "grad_norm": 0.060603220015764236, "learning_rate": 0.00019370972261137574, "loss": 1.0475, "step": 26300 }, { "epoch": 3.6474164133738602, "grad_norm": 0.11710167676210403, "learning_rate": 0.00019328943681703559, "loss": 1.0497, "step": 26400 }, { "epoch": 3.66123238463664, "grad_norm": 0.16325397789478302, "learning_rate": 0.0001928691510226954, "loss": 1.0487, "step": 26500 }, { "epoch": 3.6750483558994196, "grad_norm": 0.08937475085258484, "learning_rate": 0.00019244886522835526, "loss": 1.0468, "step": 26600 }, { "epoch": 3.6888643271621993, "grad_norm": 0.07486152648925781, "learning_rate": 0.0001920285794340151, "loss": 1.0479, "step": 26700 }, { "epoch": 3.7026802984249794, "grad_norm": 0.1263752579689026, "learning_rate": 0.00019160829363967495, "loss": 1.0449, "step": 26800 }, { "epoch": 3.716496269687759, "grad_norm": 0.11803583055734634, "learning_rate": 0.0001911880078453348, "loss": 1.0512, "step": 26900 }, { "epoch": 3.7303122409505387, "grad_norm": 0.07918773591518402, "learning_rate": 0.00019076772205099468, "loss": 1.0486, "step": 27000 }, { "epoch": 3.744128212213319, "grad_norm": 0.11923271417617798, "learning_rate": 0.00019034743625665453, "loss": 1.0465, "step": 27100 }, { "epoch": 3.7579441834760985, "grad_norm": 0.12752223014831543, "learning_rate": 0.00018992715046231437, "loss": 1.0472, "step": 27200 }, { "epoch": 3.771760154738878, "grad_norm": 0.07391146570444107, "learning_rate": 0.0001895068646679742, "loss": 1.0493, "step": 27300 }, { "epoch": 3.785576126001658, "grad_norm": 0.06606881320476532, "learning_rate": 0.00018908657887363404, "loss": 1.0485, "step": 27400 }, { "epoch": 3.7993920972644375, "grad_norm": 0.04949864745140076, "learning_rate": 0.0001886662930792939, "loss": 1.0481, "step": 27500 }, { "epoch": 3.8132080685272176, "grad_norm": 0.05234380066394806, "learning_rate": 0.00018824600728495374, "loss": 1.0476, "step": 27600 }, { "epoch": 3.8270240397899973, "grad_norm": 0.04995539411902428, "learning_rate": 0.0001878257214906136, "loss": 1.0466, "step": 27700 }, { "epoch": 3.840840011052777, "grad_norm": 0.09871330112218857, "learning_rate": 0.00018740543569627347, "loss": 1.0501, "step": 27800 }, { "epoch": 3.8546559823155566, "grad_norm": 0.06254375725984573, "learning_rate": 0.00018698514990193331, "loss": 1.0467, "step": 27900 }, { "epoch": 3.8684719535783367, "grad_norm": 0.07971449941396713, "learning_rate": 0.00018656486410759316, "loss": 1.0502, "step": 28000 }, { "epoch": 3.8822879248411164, "grad_norm": 0.12627951800823212, "learning_rate": 0.000186144578313253, "loss": 1.0446, "step": 28100 }, { "epoch": 3.896103896103896, "grad_norm": 0.08057064563035965, "learning_rate": 0.00018572429251891283, "loss": 1.0468, "step": 28200 }, { "epoch": 3.9099198673666757, "grad_norm": 0.0501413568854332, "learning_rate": 0.00018530400672457268, "loss": 1.0453, "step": 28300 }, { "epoch": 3.9237358386294554, "grad_norm": 0.09999352693557739, "learning_rate": 0.00018488372093023253, "loss": 1.0502, "step": 28400 }, { "epoch": 3.9375518098922355, "grad_norm": 0.12323564291000366, "learning_rate": 0.00018446343513589238, "loss": 1.0478, "step": 28500 }, { "epoch": 3.951367781155015, "grad_norm": 0.0877193808555603, "learning_rate": 0.00018404314934155225, "loss": 1.049, "step": 28600 }, { "epoch": 3.965183752417795, "grad_norm": 0.09397170692682266, "learning_rate": 0.0001836228635472121, "loss": 1.0474, "step": 28700 }, { "epoch": 3.978999723680575, "grad_norm": 0.09532420337200165, "learning_rate": 0.00018320257775287195, "loss": 1.0496, "step": 28800 }, { "epoch": 3.9928156949433546, "grad_norm": 0.0442403182387352, "learning_rate": 0.0001827822919585318, "loss": 1.0466, "step": 28900 }, { "epoch": 4.006631666206134, "grad_norm": 0.06309514492750168, "learning_rate": 0.00018236200616419162, "loss": 1.0479, "step": 29000 }, { "epoch": 4.020447637468914, "grad_norm": 0.06191420555114746, "learning_rate": 0.00018194172036985147, "loss": 1.0442, "step": 29100 }, { "epoch": 4.034263608731694, "grad_norm": 0.06752864271402359, "learning_rate": 0.00018152143457551132, "loss": 1.045, "step": 29200 }, { "epoch": 4.048079579994473, "grad_norm": 0.07383009046316147, "learning_rate": 0.00018110114878117117, "loss": 1.0429, "step": 29300 }, { "epoch": 4.061895551257253, "grad_norm": 0.11942852288484573, "learning_rate": 0.00018068086298683104, "loss": 1.0433, "step": 29400 }, { "epoch": 4.0757115225200335, "grad_norm": 0.0840003713965416, "learning_rate": 0.0001802605771924909, "loss": 1.0434, "step": 29500 }, { "epoch": 4.089527493782813, "grad_norm": 0.07768476754426956, "learning_rate": 0.00017984029139815074, "loss": 1.0421, "step": 29600 }, { "epoch": 4.103343465045593, "grad_norm": 0.07166603952646255, "learning_rate": 0.00017942420846175398, "loss": 1.0443, "step": 29700 }, { "epoch": 4.1171594363083726, "grad_norm": 0.07380765676498413, "learning_rate": 0.0001790039226674138, "loss": 1.0448, "step": 29800 }, { "epoch": 4.130975407571152, "grad_norm": 0.1263025552034378, "learning_rate": 0.00017858363687307365, "loss": 1.0437, "step": 29900 }, { "epoch": 4.144791378833932, "grad_norm": 0.09632286429405212, "learning_rate": 0.00017816335107873353, "loss": 1.0439, "step": 30000 }, { "epoch": 4.144791378833932, "eval_accuracy": 0.5233148259844476, "eval_loss": 1.0439139604568481, "eval_runtime": 787.8404, "eval_samples_per_second": 261.298, "eval_steps_per_second": 8.167, "step": 30000 }, { "epoch": 4.158607350096712, "grad_norm": 0.09395026415586472, "learning_rate": 0.00017774306528439338, "loss": 1.0447, "step": 30100 }, { "epoch": 4.172423321359491, "grad_norm": 0.07320912927389145, "learning_rate": 0.00017732277949005323, "loss": 1.0477, "step": 30200 }, { "epoch": 4.186239292622272, "grad_norm": 0.05703623965382576, "learning_rate": 0.00017690249369571308, "loss": 1.0443, "step": 30300 }, { "epoch": 4.2000552638850515, "grad_norm": 0.04885410889983177, "learning_rate": 0.00017648220790137292, "loss": 1.0467, "step": 30400 }, { "epoch": 4.213871235147831, "grad_norm": 0.10649748146533966, "learning_rate": 0.00017606192210703277, "loss": 1.0448, "step": 30500 }, { "epoch": 4.227687206410611, "grad_norm": 0.05844441428780556, "learning_rate": 0.0001756416363126926, "loss": 1.044, "step": 30600 }, { "epoch": 4.2415031776733905, "grad_norm": 0.07287675887346268, "learning_rate": 0.00017522135051835244, "loss": 1.0428, "step": 30700 }, { "epoch": 4.25531914893617, "grad_norm": 0.05190150439739227, "learning_rate": 0.00017480106472401232, "loss": 1.0413, "step": 30800 }, { "epoch": 4.26913512019895, "grad_norm": 0.06985218822956085, "learning_rate": 0.00017438077892967217, "loss": 1.0455, "step": 30900 }, { "epoch": 4.2829510914617295, "grad_norm": 0.06930764764547348, "learning_rate": 0.00017396049313533202, "loss": 1.0444, "step": 31000 }, { "epoch": 4.296767062724509, "grad_norm": 0.07905230671167374, "learning_rate": 0.00017354020734099186, "loss": 1.0445, "step": 31100 }, { "epoch": 4.31058303398729, "grad_norm": 0.04994554817676544, "learning_rate": 0.0001731199215466517, "loss": 1.0432, "step": 31200 }, { "epoch": 4.324399005250069, "grad_norm": 0.08036911487579346, "learning_rate": 0.00017269963575231156, "loss": 1.0424, "step": 31300 }, { "epoch": 4.338214976512849, "grad_norm": 0.07251475006341934, "learning_rate": 0.00017227934995797138, "loss": 1.0465, "step": 31400 }, { "epoch": 4.352030947775629, "grad_norm": 0.09622683376073837, "learning_rate": 0.00017185906416363123, "loss": 1.0441, "step": 31500 }, { "epoch": 4.365846919038408, "grad_norm": 0.07545050978660583, "learning_rate": 0.0001714387783692911, "loss": 1.0423, "step": 31600 }, { "epoch": 4.379662890301188, "grad_norm": 0.07171428948640823, "learning_rate": 0.00017102269543289435, "loss": 1.0434, "step": 31700 }, { "epoch": 4.393478861563968, "grad_norm": 0.06658755987882614, "learning_rate": 0.0001706024096385542, "loss": 1.0415, "step": 31800 }, { "epoch": 4.407294832826747, "grad_norm": 0.10734014213085175, "learning_rate": 0.00017018212384421405, "loss": 1.0406, "step": 31900 }, { "epoch": 4.421110804089528, "grad_norm": 0.06358776986598969, "learning_rate": 0.0001697618380498739, "loss": 1.0405, "step": 32000 }, { "epoch": 4.434926775352308, "grad_norm": 0.06078578904271126, "learning_rate": 0.00016934155225553377, "loss": 1.0458, "step": 32100 }, { "epoch": 4.448742746615087, "grad_norm": 0.09674441814422607, "learning_rate": 0.000168925469319137, "loss": 1.0433, "step": 32200 }, { "epoch": 4.462558717877867, "grad_norm": 0.11840452253818512, "learning_rate": 0.00016850518352479684, "loss": 1.0448, "step": 32300 }, { "epoch": 4.476374689140647, "grad_norm": 0.08742488920688629, "learning_rate": 0.0001680848977304567, "loss": 1.0409, "step": 32400 }, { "epoch": 4.490190660403426, "grad_norm": 0.09082327783107758, "learning_rate": 0.00016766461193611654, "loss": 1.0432, "step": 32500 }, { "epoch": 4.504006631666206, "grad_norm": 0.06259270012378693, "learning_rate": 0.0001672443261417764, "loss": 1.0406, "step": 32600 }, { "epoch": 4.517822602928986, "grad_norm": 0.06466669589281082, "learning_rate": 0.00016682404034743626, "loss": 1.0404, "step": 32700 }, { "epoch": 4.531638574191765, "grad_norm": 0.07167832553386688, "learning_rate": 0.0001664037545530961, "loss": 1.0457, "step": 32800 }, { "epoch": 4.545454545454545, "grad_norm": 0.055970191955566406, "learning_rate": 0.00016598346875875596, "loss": 1.0433, "step": 32900 }, { "epoch": 4.5592705167173255, "grad_norm": 0.05038364604115486, "learning_rate": 0.00016556318296441578, "loss": 1.0414, "step": 33000 }, { "epoch": 4.573086487980105, "grad_norm": 0.11647244542837143, "learning_rate": 0.00016514289717007563, "loss": 1.0408, "step": 33100 }, { "epoch": 4.586902459242885, "grad_norm": 0.08881094306707382, "learning_rate": 0.00016472261137573548, "loss": 1.0468, "step": 33200 }, { "epoch": 4.6007184305056645, "grad_norm": 0.0706004872918129, "learning_rate": 0.00016430232558139533, "loss": 1.0433, "step": 33300 }, { "epoch": 4.614534401768444, "grad_norm": 0.07594550400972366, "learning_rate": 0.00016388203978705518, "loss": 1.0401, "step": 33400 }, { "epoch": 4.628350373031224, "grad_norm": 0.06709697842597961, "learning_rate": 0.00016346175399271505, "loss": 1.0406, "step": 33500 }, { "epoch": 4.6421663442940035, "grad_norm": 0.055218733847141266, "learning_rate": 0.0001630414681983749, "loss": 1.0439, "step": 33600 }, { "epoch": 4.655982315556784, "grad_norm": 0.09484557062387466, "learning_rate": 0.00016262118240403475, "loss": 1.0445, "step": 33700 }, { "epoch": 4.669798286819564, "grad_norm": 0.08181110769510269, "learning_rate": 0.00016220089660969457, "loss": 1.0404, "step": 33800 }, { "epoch": 4.683614258082343, "grad_norm": 0.07101566344499588, "learning_rate": 0.00016178061081535442, "loss": 1.0418, "step": 33900 }, { "epoch": 4.697430229345123, "grad_norm": 0.07521411031484604, "learning_rate": 0.00016136032502101427, "loss": 1.0413, "step": 34000 }, { "epoch": 4.711246200607903, "grad_norm": 0.06438640505075455, "learning_rate": 0.00016094003922667412, "loss": 1.0413, "step": 34100 }, { "epoch": 4.7250621718706824, "grad_norm": 0.0852956548333168, "learning_rate": 0.00016051975343233396, "loss": 1.0411, "step": 34200 }, { "epoch": 4.738878143133462, "grad_norm": 0.041669171303510666, "learning_rate": 0.00016009946763799384, "loss": 1.043, "step": 34300 }, { "epoch": 4.752694114396242, "grad_norm": 0.07866424322128296, "learning_rate": 0.0001596791818436537, "loss": 1.0416, "step": 34400 }, { "epoch": 4.7665100856590215, "grad_norm": 0.06820093840360641, "learning_rate": 0.00015925889604931354, "loss": 1.0419, "step": 34500 }, { "epoch": 4.780326056921801, "grad_norm": 0.08769433945417404, "learning_rate": 0.00015883861025497336, "loss": 1.0436, "step": 34600 }, { "epoch": 4.794142028184582, "grad_norm": 0.11472765356302261, "learning_rate": 0.0001584183244606332, "loss": 1.0448, "step": 34700 }, { "epoch": 4.807957999447361, "grad_norm": 0.10286398231983185, "learning_rate": 0.00015799803866629305, "loss": 1.0396, "step": 34800 }, { "epoch": 4.821773970710141, "grad_norm": 0.08412828296422958, "learning_rate": 0.0001575777528719529, "loss": 1.0432, "step": 34900 }, { "epoch": 4.835589941972921, "grad_norm": 0.06536369025707245, "learning_rate": 0.00015715746707761275, "loss": 1.0425, "step": 35000 }, { "epoch": 4.835589941972921, "eval_accuracy": 0.5253784900927014, "eval_loss": 1.0407328605651855, "eval_runtime": 804.3369, "eval_samples_per_second": 255.939, "eval_steps_per_second": 7.999, "step": 35000 }, { "epoch": 4.8494059132357, "grad_norm": 0.05366332083940506, "learning_rate": 0.00015673718128327263, "loss": 1.0401, "step": 35100 }, { "epoch": 4.86322188449848, "grad_norm": 0.05627182498574257, "learning_rate": 0.00015631689548893248, "loss": 1.0413, "step": 35200 }, { "epoch": 4.87703785576126, "grad_norm": 0.06880544126033783, "learning_rate": 0.00015589660969459232, "loss": 1.0399, "step": 35300 }, { "epoch": 4.89085382702404, "grad_norm": 0.06326279044151306, "learning_rate": 0.00015547632390025215, "loss": 1.0424, "step": 35400 }, { "epoch": 4.90466979828682, "grad_norm": 0.050615083426237106, "learning_rate": 0.000155056038105912, "loss": 1.0419, "step": 35500 }, { "epoch": 4.9184857695496, "grad_norm": 0.09092865139245987, "learning_rate": 0.00015463575231157184, "loss": 1.0417, "step": 35600 }, { "epoch": 4.932301740812379, "grad_norm": 0.10828616470098495, "learning_rate": 0.0001542154665172317, "loss": 1.0461, "step": 35700 }, { "epoch": 4.946117712075159, "grad_norm": 0.10398013889789581, "learning_rate": 0.00015379518072289154, "loss": 1.0402, "step": 35800 }, { "epoch": 4.959933683337939, "grad_norm": 0.060978490859270096, "learning_rate": 0.00015337489492855142, "loss": 1.0428, "step": 35900 }, { "epoch": 4.973749654600718, "grad_norm": 0.09474412351846695, "learning_rate": 0.00015295460913421126, "loss": 1.0426, "step": 36000 }, { "epoch": 4.987565625863498, "grad_norm": 0.055337630212306976, "learning_rate": 0.0001525343233398711, "loss": 1.0424, "step": 36100 }, { "epoch": 5.001381597126278, "grad_norm": 0.062282662838697433, "learning_rate": 0.00015211824040347433, "loss": 1.0408, "step": 36200 }, { "epoch": 5.015197568389058, "grad_norm": 0.08418793976306915, "learning_rate": 0.00015169795460913418, "loss": 1.0423, "step": 36300 }, { "epoch": 5.029013539651838, "grad_norm": 0.056806761771440506, "learning_rate": 0.00015127766881479403, "loss": 1.0397, "step": 36400 }, { "epoch": 5.0428295109146175, "grad_norm": 0.050782449543476105, "learning_rate": 0.0001508573830204539, "loss": 1.0397, "step": 36500 }, { "epoch": 5.056645482177397, "grad_norm": 0.04436805471777916, "learning_rate": 0.00015043709722611375, "loss": 1.0372, "step": 36600 }, { "epoch": 5.070461453440177, "grad_norm": 0.056697145104408264, "learning_rate": 0.0001500168114317736, "loss": 1.0396, "step": 36700 }, { "epoch": 5.0842774247029565, "grad_norm": 0.0936078131198883, "learning_rate": 0.00014959652563743342, "loss": 1.0366, "step": 36800 }, { "epoch": 5.098093395965736, "grad_norm": 0.058340467512607574, "learning_rate": 0.0001491762398430933, "loss": 1.038, "step": 36900 }, { "epoch": 5.111909367228516, "grad_norm": 0.07920562475919724, "learning_rate": 0.00014875595404875315, "loss": 1.0389, "step": 37000 }, { "epoch": 5.1257253384912955, "grad_norm": 0.054546140134334564, "learning_rate": 0.000148335668254413, "loss": 1.0352, "step": 37100 }, { "epoch": 5.139541309754076, "grad_norm": 0.0779619961977005, "learning_rate": 0.00014791538246007282, "loss": 1.0362, "step": 37200 }, { "epoch": 5.153357281016856, "grad_norm": 0.06077539920806885, "learning_rate": 0.0001474950966657327, "loss": 1.0395, "step": 37300 }, { "epoch": 5.167173252279635, "grad_norm": 0.07015964388847351, "learning_rate": 0.00014707481087139254, "loss": 1.0378, "step": 37400 }, { "epoch": 5.180989223542415, "grad_norm": 0.07821048051118851, "learning_rate": 0.0001466545250770524, "loss": 1.0358, "step": 37500 }, { "epoch": 5.194805194805195, "grad_norm": 0.06446918845176697, "learning_rate": 0.0001462342392827122, "loss": 1.0401, "step": 37600 }, { "epoch": 5.208621166067974, "grad_norm": 0.0754179060459137, "learning_rate": 0.0001458139534883721, "loss": 1.0372, "step": 37700 }, { "epoch": 5.222437137330754, "grad_norm": 0.06225774064660072, "learning_rate": 0.00014539366769403194, "loss": 1.0396, "step": 37800 }, { "epoch": 5.236253108593534, "grad_norm": 0.09567879885435104, "learning_rate": 0.00014497338189969178, "loss": 1.0427, "step": 37900 }, { "epoch": 5.250069079856313, "grad_norm": 0.0810612216591835, "learning_rate": 0.00014455309610535163, "loss": 1.0368, "step": 38000 }, { "epoch": 5.263885051119094, "grad_norm": 0.058250732719898224, "learning_rate": 0.00014413281031101148, "loss": 1.039, "step": 38100 }, { "epoch": 5.277701022381874, "grad_norm": 0.07354842871427536, "learning_rate": 0.00014371252451667133, "loss": 1.0393, "step": 38200 }, { "epoch": 5.291516993644653, "grad_norm": 0.04756517335772514, "learning_rate": 0.00014329223872233118, "loss": 1.0369, "step": 38300 }, { "epoch": 5.305332964907433, "grad_norm": 0.05551883205771446, "learning_rate": 0.00014287195292799103, "loss": 1.038, "step": 38400 }, { "epoch": 5.319148936170213, "grad_norm": 0.05476289987564087, "learning_rate": 0.00014245166713365088, "loss": 1.0391, "step": 38500 }, { "epoch": 5.332964907432992, "grad_norm": 0.041929882019758224, "learning_rate": 0.00014203138133931072, "loss": 1.0377, "step": 38600 }, { "epoch": 5.346780878695772, "grad_norm": 0.05916072428226471, "learning_rate": 0.00014161109554497057, "loss": 1.0417, "step": 38700 }, { "epoch": 5.360596849958552, "grad_norm": 0.0609772689640522, "learning_rate": 0.00014119080975063042, "loss": 1.0386, "step": 38800 }, { "epoch": 5.374412821221332, "grad_norm": 0.06430498510599136, "learning_rate": 0.00014077052395629027, "loss": 1.0397, "step": 38900 }, { "epoch": 5.388228792484112, "grad_norm": 0.07042800635099411, "learning_rate": 0.00014035023816195012, "loss": 1.038, "step": 39000 }, { "epoch": 5.402044763746892, "grad_norm": 0.05623612925410271, "learning_rate": 0.00013992995236760997, "loss": 1.0405, "step": 39100 }, { "epoch": 5.415860735009671, "grad_norm": 0.04936366528272629, "learning_rate": 0.00013950966657326982, "loss": 1.0404, "step": 39200 }, { "epoch": 5.429676706272451, "grad_norm": 0.05738508701324463, "learning_rate": 0.00013908938077892966, "loss": 1.0364, "step": 39300 }, { "epoch": 5.443492677535231, "grad_norm": 0.09567712992429733, "learning_rate": 0.0001386690949845895, "loss": 1.0381, "step": 39400 }, { "epoch": 5.45730864879801, "grad_norm": 0.07306545972824097, "learning_rate": 0.00013824880919024936, "loss": 1.0394, "step": 39500 }, { "epoch": 5.47112462006079, "grad_norm": 0.060108475387096405, "learning_rate": 0.0001378285233959092, "loss": 1.0379, "step": 39600 }, { "epoch": 5.48494059132357, "grad_norm": 0.08150669932365417, "learning_rate": 0.00013740823760156906, "loss": 1.0391, "step": 39700 }, { "epoch": 5.49875656258635, "grad_norm": 0.06265643239021301, "learning_rate": 0.0001369879518072289, "loss": 1.0419, "step": 39800 }, { "epoch": 5.51257253384913, "grad_norm": 0.09023050218820572, "learning_rate": 0.00013656766601288876, "loss": 1.0374, "step": 39900 }, { "epoch": 5.5263885051119095, "grad_norm": 0.06600885838270187, "learning_rate": 0.0001361473802185486, "loss": 1.0365, "step": 40000 }, { "epoch": 5.5263885051119095, "eval_accuracy": 0.52706640122358, "eval_loss": 1.0380040407180786, "eval_runtime": 773.4583, "eval_samples_per_second": 266.157, "eval_steps_per_second": 8.318, "step": 40000 }, { "epoch": 5.540204476374689, "grad_norm": 0.07041644304990768, "learning_rate": 0.00013572709442420845, "loss": 1.038, "step": 40100 }, { "epoch": 5.554020447637469, "grad_norm": 0.0819341391324997, "learning_rate": 0.0001353110114878117, "loss": 1.0383, "step": 40200 }, { "epoch": 5.5678364189002485, "grad_norm": 0.04390214383602142, "learning_rate": 0.00013489072569347155, "loss": 1.0381, "step": 40300 }, { "epoch": 5.581652390163028, "grad_norm": 0.0681944414973259, "learning_rate": 0.0001344704398991314, "loss": 1.0368, "step": 40400 }, { "epoch": 5.595468361425809, "grad_norm": 0.0888848677277565, "learning_rate": 0.00013405015410479124, "loss": 1.0369, "step": 40500 }, { "epoch": 5.609284332688588, "grad_norm": 0.07275230437517166, "learning_rate": 0.0001336298683104511, "loss": 1.0353, "step": 40600 }, { "epoch": 5.623100303951368, "grad_norm": 0.10200846940279007, "learning_rate": 0.00013320958251611094, "loss": 1.0381, "step": 40700 }, { "epoch": 5.636916275214148, "grad_norm": 0.056480832397937775, "learning_rate": 0.0001327892967217708, "loss": 1.0383, "step": 40800 }, { "epoch": 5.650732246476927, "grad_norm": 0.0845484584569931, "learning_rate": 0.00013236901092743064, "loss": 1.0385, "step": 40900 }, { "epoch": 5.664548217739707, "grad_norm": 0.05990500748157501, "learning_rate": 0.0001319487251330905, "loss": 1.0381, "step": 41000 }, { "epoch": 5.678364189002487, "grad_norm": 0.04566818103194237, "learning_rate": 0.00013152843933875034, "loss": 1.0409, "step": 41100 }, { "epoch": 5.692180160265266, "grad_norm": 0.05529521405696869, "learning_rate": 0.00013110815354441018, "loss": 1.039, "step": 41200 }, { "epoch": 5.705996131528046, "grad_norm": 0.08812158554792404, "learning_rate": 0.00013068786775007003, "loss": 1.0393, "step": 41300 }, { "epoch": 5.719812102790826, "grad_norm": 0.0714721605181694, "learning_rate": 0.00013026758195572988, "loss": 1.0365, "step": 41400 }, { "epoch": 5.733628074053606, "grad_norm": 0.050889432430267334, "learning_rate": 0.00012984729616138973, "loss": 1.0399, "step": 41500 }, { "epoch": 5.747444045316386, "grad_norm": 0.05863107368350029, "learning_rate": 0.00012942701036704958, "loss": 1.0401, "step": 41600 }, { "epoch": 5.761260016579166, "grad_norm": 0.05279000476002693, "learning_rate": 0.00012900672457270943, "loss": 1.0368, "step": 41700 }, { "epoch": 5.775075987841945, "grad_norm": 0.06430874019861221, "learning_rate": 0.00012858643877836928, "loss": 1.0347, "step": 41800 }, { "epoch": 5.788891959104725, "grad_norm": 0.1187288910150528, "learning_rate": 0.00012816615298402912, "loss": 1.0372, "step": 41900 }, { "epoch": 5.802707930367505, "grad_norm": 0.05984746664762497, "learning_rate": 0.00012774586718968897, "loss": 1.036, "step": 42000 }, { "epoch": 5.816523901630284, "grad_norm": 0.047202371060848236, "learning_rate": 0.00012732558139534882, "loss": 1.0341, "step": 42100 }, { "epoch": 5.830339872893065, "grad_norm": 0.0888022631406784, "learning_rate": 0.00012690949845895207, "loss": 1.0358, "step": 42200 }, { "epoch": 5.8441558441558445, "grad_norm": 0.071753591299057, "learning_rate": 0.00012648921266461191, "loss": 1.0356, "step": 42300 }, { "epoch": 5.857971815418624, "grad_norm": 0.06311481446027756, "learning_rate": 0.0001260689268702718, "loss": 1.0381, "step": 42400 }, { "epoch": 5.871787786681404, "grad_norm": 0.05733519420027733, "learning_rate": 0.0001256486410759316, "loss": 1.0366, "step": 42500 }, { "epoch": 5.885603757944184, "grad_norm": 0.05296749621629715, "learning_rate": 0.00012522835528159146, "loss": 1.0391, "step": 42600 }, { "epoch": 5.899419729206963, "grad_norm": 0.05728083476424217, "learning_rate": 0.0001248080694872513, "loss": 1.0393, "step": 42700 }, { "epoch": 5.913235700469743, "grad_norm": 0.10918726772069931, "learning_rate": 0.00012438778369291118, "loss": 1.0375, "step": 42800 }, { "epoch": 5.927051671732523, "grad_norm": 0.043641045689582825, "learning_rate": 0.000123967497898571, "loss": 1.0342, "step": 42900 }, { "epoch": 5.940867642995302, "grad_norm": 0.07793564349412918, "learning_rate": 0.00012354721210423085, "loss": 1.037, "step": 43000 }, { "epoch": 5.954683614258082, "grad_norm": 0.10596407949924469, "learning_rate": 0.0001231269263098907, "loss": 1.0361, "step": 43100 }, { "epoch": 5.9684995855208625, "grad_norm": 0.05018968880176544, "learning_rate": 0.00012270664051555058, "loss": 1.0352, "step": 43200 }, { "epoch": 5.982315556783642, "grad_norm": 0.06663347035646439, "learning_rate": 0.0001222863547212104, "loss": 1.0379, "step": 43300 }, { "epoch": 5.996131528046422, "grad_norm": 0.05061174929141998, "learning_rate": 0.00012186606892687026, "loss": 1.0378, "step": 43400 }, { "epoch": 6.0099474993092015, "grad_norm": 0.07496211677789688, "learning_rate": 0.00012144578313253011, "loss": 1.0357, "step": 43500 }, { "epoch": 6.023763470571981, "grad_norm": 0.058973684906959534, "learning_rate": 0.00012102549733818996, "loss": 1.0336, "step": 43600 }, { "epoch": 6.037579441834761, "grad_norm": 0.07304850965738297, "learning_rate": 0.0001206052115438498, "loss": 1.0366, "step": 43700 }, { "epoch": 6.0513954130975405, "grad_norm": 0.05964922904968262, "learning_rate": 0.00012018492574950966, "loss": 1.0358, "step": 43800 }, { "epoch": 6.06521138436032, "grad_norm": 0.10107408463954926, "learning_rate": 0.0001197646399551695, "loss": 1.0363, "step": 43900 }, { "epoch": 6.079027355623101, "grad_norm": 0.05830320343375206, "learning_rate": 0.00011934435416082935, "loss": 1.0374, "step": 44000 }, { "epoch": 6.09284332688588, "grad_norm": 0.06493101269006729, "learning_rate": 0.00011892406836648919, "loss": 1.0358, "step": 44100 }, { "epoch": 6.10665929814866, "grad_norm": 0.06381756067276001, "learning_rate": 0.00011850798543009245, "loss": 1.0345, "step": 44200 }, { "epoch": 6.12047526941144, "grad_norm": 0.057328786700963974, "learning_rate": 0.0001180876996357523, "loss": 1.0347, "step": 44300 }, { "epoch": 6.134291240674219, "grad_norm": 0.09036822617053986, "learning_rate": 0.00011766741384141216, "loss": 1.0352, "step": 44400 }, { "epoch": 6.148107211936999, "grad_norm": 0.05485937371850014, "learning_rate": 0.000117247128047072, "loss": 1.0371, "step": 44500 }, { "epoch": 6.161923183199779, "grad_norm": 0.06304465979337692, "learning_rate": 0.00011682684225273184, "loss": 1.0302, "step": 44600 }, { "epoch": 6.175739154462558, "grad_norm": 0.045126065611839294, "learning_rate": 0.0001164065564583917, "loss": 1.0338, "step": 44700 }, { "epoch": 6.189555125725338, "grad_norm": 0.06636038422584534, "learning_rate": 0.00011598627066405155, "loss": 1.0353, "step": 44800 }, { "epoch": 6.203371096988119, "grad_norm": 0.05977385491132736, "learning_rate": 0.00011556598486971139, "loss": 1.0346, "step": 44900 }, { "epoch": 6.217187068250898, "grad_norm": 0.07459376752376556, "learning_rate": 0.00011514569907537124, "loss": 1.0325, "step": 45000 }, { "epoch": 6.217187068250898, "eval_accuracy": 0.5284276106869993, "eval_loss": 1.0360603332519531, "eval_runtime": 770.702, "eval_samples_per_second": 267.108, "eval_steps_per_second": 8.348, "step": 45000 }, { "epoch": 6.231003039513678, "grad_norm": 0.050757069140672684, "learning_rate": 0.0001147254132810311, "loss": 1.0337, "step": 45100 }, { "epoch": 6.244819010776458, "grad_norm": 0.065644271671772, "learning_rate": 0.00011430512748669095, "loss": 1.035, "step": 45200 }, { "epoch": 6.258634982039237, "grad_norm": 0.06008651480078697, "learning_rate": 0.00011388484169235078, "loss": 1.0323, "step": 45300 }, { "epoch": 6.272450953302017, "grad_norm": 0.050868868827819824, "learning_rate": 0.00011346455589801063, "loss": 1.0341, "step": 45400 }, { "epoch": 6.286266924564797, "grad_norm": 0.0535401850938797, "learning_rate": 0.00011304427010367049, "loss": 1.0349, "step": 45500 }, { "epoch": 6.300082895827576, "grad_norm": 0.07083383947610855, "learning_rate": 0.00011262398430933034, "loss": 1.0327, "step": 45600 }, { "epoch": 6.313898867090357, "grad_norm": 0.06998474150896072, "learning_rate": 0.00011220369851499018, "loss": 1.035, "step": 45700 }, { "epoch": 6.3277148383531365, "grad_norm": 0.06696050614118576, "learning_rate": 0.00011178341272065002, "loss": 1.0342, "step": 45800 }, { "epoch": 6.341530809615916, "grad_norm": 0.050143785774707794, "learning_rate": 0.00011136312692630989, "loss": 1.0342, "step": 45900 }, { "epoch": 6.355346780878696, "grad_norm": 0.066258005797863, "learning_rate": 0.00011094284113196974, "loss": 1.0368, "step": 46000 }, { "epoch": 6.3691627521414755, "grad_norm": 0.057613175362348557, "learning_rate": 0.00011052255533762957, "loss": 1.0357, "step": 46100 }, { "epoch": 6.382978723404255, "grad_norm": 0.07405593246221542, "learning_rate": 0.00011010647240123283, "loss": 1.033, "step": 46200 }, { "epoch": 6.396794694667035, "grad_norm": 0.07005150616168976, "learning_rate": 0.00010968618660689268, "loss": 1.0329, "step": 46300 }, { "epoch": 6.4106106659298145, "grad_norm": 0.057546067982912064, "learning_rate": 0.00010926590081255253, "loss": 1.033, "step": 46400 }, { "epoch": 6.424426637192594, "grad_norm": 0.08016248792409897, "learning_rate": 0.00010884561501821236, "loss": 1.0389, "step": 46500 }, { "epoch": 6.438242608455375, "grad_norm": 0.08346617966890335, "learning_rate": 0.00010842532922387222, "loss": 1.0332, "step": 46600 }, { "epoch": 6.452058579718154, "grad_norm": 0.048157453536987305, "learning_rate": 0.00010800504342953207, "loss": 1.0342, "step": 46700 }, { "epoch": 6.465874550980934, "grad_norm": 0.06816009432077408, "learning_rate": 0.00010758475763519192, "loss": 1.0357, "step": 46800 }, { "epoch": 6.479690522243714, "grad_norm": 0.05210613086819649, "learning_rate": 0.00010716447184085176, "loss": 1.0345, "step": 46900 }, { "epoch": 6.4935064935064934, "grad_norm": 0.08138227462768555, "learning_rate": 0.00010674418604651162, "loss": 1.035, "step": 47000 }, { "epoch": 6.507322464769273, "grad_norm": 0.07494477927684784, "learning_rate": 0.00010632390025217147, "loss": 1.0361, "step": 47100 }, { "epoch": 6.521138436032053, "grad_norm": 0.07473413646221161, "learning_rate": 0.00010590361445783132, "loss": 1.0339, "step": 47200 }, { "epoch": 6.5349544072948325, "grad_norm": 0.07200802862644196, "learning_rate": 0.00010548332866349115, "loss": 1.0333, "step": 47300 }, { "epoch": 6.548770378557613, "grad_norm": 0.06346756964921951, "learning_rate": 0.00010506304286915101, "loss": 1.0345, "step": 47400 }, { "epoch": 6.562586349820393, "grad_norm": 0.06382066756486893, "learning_rate": 0.00010464275707481086, "loss": 1.0352, "step": 47500 }, { "epoch": 6.576402321083172, "grad_norm": 0.1000475063920021, "learning_rate": 0.00010422247128047071, "loss": 1.0344, "step": 47600 }, { "epoch": 6.590218292345952, "grad_norm": 0.06456384807825089, "learning_rate": 0.00010380218548613057, "loss": 1.0356, "step": 47700 }, { "epoch": 6.604034263608732, "grad_norm": 0.052929963916540146, "learning_rate": 0.0001033818996917904, "loss": 1.0343, "step": 47800 }, { "epoch": 6.617850234871511, "grad_norm": 0.07275223731994629, "learning_rate": 0.00010296161389745025, "loss": 1.033, "step": 47900 }, { "epoch": 6.631666206134291, "grad_norm": 0.060610584914684296, "learning_rate": 0.0001025413281031101, "loss": 1.0334, "step": 48000 }, { "epoch": 6.645482177397071, "grad_norm": 0.0514766089618206, "learning_rate": 0.00010212104230876997, "loss": 1.0351, "step": 48100 }, { "epoch": 6.65929814865985, "grad_norm": 0.08950326591730118, "learning_rate": 0.0001017049593723732, "loss": 1.0341, "step": 48200 }, { "epoch": 6.673114119922631, "grad_norm": 0.052268847823143005, "learning_rate": 0.00010128467357803306, "loss": 1.0342, "step": 48300 }, { "epoch": 6.686930091185411, "grad_norm": 0.059182267636060715, "learning_rate": 0.00010086438778369291, "loss": 1.0303, "step": 48400 }, { "epoch": 6.70074606244819, "grad_norm": 0.06220945715904236, "learning_rate": 0.00010044410198935274, "loss": 1.032, "step": 48500 }, { "epoch": 6.71456203371097, "grad_norm": 0.0486241914331913, "learning_rate": 0.00010002381619501259, "loss": 1.0338, "step": 48600 }, { "epoch": 6.72837800497375, "grad_norm": 0.04813262075185776, "learning_rate": 9.960353040067245e-05, "loss": 1.0344, "step": 48700 }, { "epoch": 6.742193976236529, "grad_norm": 0.04981222748756409, "learning_rate": 9.91832446063323e-05, "loss": 1.0347, "step": 48800 }, { "epoch": 6.756009947499309, "grad_norm": 0.050560541450977325, "learning_rate": 9.876295881199214e-05, "loss": 1.0338, "step": 48900 }, { "epoch": 6.769825918762089, "grad_norm": 0.05338674411177635, "learning_rate": 9.834267301765199e-05, "loss": 1.0369, "step": 49000 }, { "epoch": 6.783641890024869, "grad_norm": 0.042156435549259186, "learning_rate": 9.792238722331185e-05, "loss": 1.0345, "step": 49100 }, { "epoch": 6.797457861287649, "grad_norm": 0.0622396394610405, "learning_rate": 9.75021014289717e-05, "loss": 1.0321, "step": 49200 }, { "epoch": 6.8112738325504285, "grad_norm": 0.08523661643266678, "learning_rate": 9.708181563463155e-05, "loss": 1.0317, "step": 49300 }, { "epoch": 6.825089803813208, "grad_norm": 0.055176641792058945, "learning_rate": 9.666152984029138e-05, "loss": 1.0368, "step": 49400 }, { "epoch": 6.838905775075988, "grad_norm": 0.07358380407094955, "learning_rate": 9.624124404595124e-05, "loss": 1.0318, "step": 49500 }, { "epoch": 6.8527217463387675, "grad_norm": 0.055568769574165344, "learning_rate": 9.582095825161109e-05, "loss": 1.0343, "step": 49600 }, { "epoch": 6.866537717601547, "grad_norm": 0.04249552637338638, "learning_rate": 9.540067245727094e-05, "loss": 1.0331, "step": 49700 }, { "epoch": 6.880353688864327, "grad_norm": 0.05274058133363724, "learning_rate": 9.498038666293077e-05, "loss": 1.0351, "step": 49800 }, { "epoch": 6.8941696601271065, "grad_norm": 0.04792112484574318, "learning_rate": 9.456010086859064e-05, "loss": 1.0333, "step": 49900 }, { "epoch": 6.907985631389887, "grad_norm": 0.05513302981853485, "learning_rate": 9.413981507425049e-05, "loss": 1.0322, "step": 50000 }, { "epoch": 6.907985631389887, "eval_accuracy": 0.5296076152096916, "eval_loss": 1.0341060161590576, "eval_runtime": 725.8939, "eval_samples_per_second": 283.597, "eval_steps_per_second": 8.864, "step": 50000 }, { "epoch": 6.921801602652667, "grad_norm": 0.05296773836016655, "learning_rate": 9.371952927991033e-05, "loss": 1.031, "step": 50100 }, { "epoch": 6.935617573915446, "grad_norm": 0.062248583883047104, "learning_rate": 9.330344634351358e-05, "loss": 1.0341, "step": 50200 }, { "epoch": 6.949433545178226, "grad_norm": 0.07751675695180893, "learning_rate": 9.288316054917343e-05, "loss": 1.0352, "step": 50300 }, { "epoch": 6.963249516441006, "grad_norm": 0.04984898492693901, "learning_rate": 9.246287475483328e-05, "loss": 1.0302, "step": 50400 }, { "epoch": 6.977065487703785, "grad_norm": 0.04315504804253578, "learning_rate": 9.204258896049314e-05, "loss": 1.0327, "step": 50500 }, { "epoch": 6.990881458966565, "grad_norm": 0.053620435297489166, "learning_rate": 9.162230316615297e-05, "loss": 1.0328, "step": 50600 }, { "epoch": 7.004697430229345, "grad_norm": 0.04611975699663162, "learning_rate": 9.120201737181282e-05, "loss": 1.0336, "step": 50700 }, { "epoch": 7.018513401492125, "grad_norm": 0.04269848018884659, "learning_rate": 9.078173157747267e-05, "loss": 1.0282, "step": 50800 }, { "epoch": 7.032329372754905, "grad_norm": 0.055365532636642456, "learning_rate": 9.036144578313253e-05, "loss": 1.0339, "step": 50900 }, { "epoch": 7.046145344017685, "grad_norm": 0.06129321828484535, "learning_rate": 8.994115998879237e-05, "loss": 1.0304, "step": 51000 }, { "epoch": 7.059961315280464, "grad_norm": 0.06094348803162575, "learning_rate": 8.952507705239563e-05, "loss": 1.0288, "step": 51100 }, { "epoch": 7.073777286543244, "grad_norm": 0.048849135637283325, "learning_rate": 8.910479125805548e-05, "loss": 1.0322, "step": 51200 }, { "epoch": 7.087593257806024, "grad_norm": 0.05081125721335411, "learning_rate": 8.868450546371531e-05, "loss": 1.0303, "step": 51300 }, { "epoch": 7.101409229068803, "grad_norm": 0.07727497071027756, "learning_rate": 8.826421966937516e-05, "loss": 1.03, "step": 51400 }, { "epoch": 7.115225200331583, "grad_norm": 0.06357153505086899, "learning_rate": 8.784393387503502e-05, "loss": 1.0342, "step": 51500 }, { "epoch": 7.129041171594363, "grad_norm": 0.05598052963614464, "learning_rate": 8.742364808069487e-05, "loss": 1.0312, "step": 51600 }, { "epoch": 7.142857142857143, "grad_norm": 0.06753697246313095, "learning_rate": 8.70033622863547e-05, "loss": 1.0306, "step": 51700 }, { "epoch": 7.156673114119923, "grad_norm": 0.06586912274360657, "learning_rate": 8.658307649201455e-05, "loss": 1.0311, "step": 51800 }, { "epoch": 7.170489085382703, "grad_norm": 0.10361455380916595, "learning_rate": 8.616279069767442e-05, "loss": 1.0326, "step": 51900 }, { "epoch": 7.184305056645482, "grad_norm": 0.09442713856697083, "learning_rate": 8.574250490333426e-05, "loss": 1.0339, "step": 52000 }, { "epoch": 7.198121027908262, "grad_norm": 0.08114325255155563, "learning_rate": 8.532221910899411e-05, "loss": 1.0335, "step": 52100 }, { "epoch": 7.211936999171042, "grad_norm": 0.054252710193395615, "learning_rate": 8.490193331465395e-05, "loss": 1.0316, "step": 52200 }, { "epoch": 7.225752970433821, "grad_norm": 0.059643086045980453, "learning_rate": 8.448164752031381e-05, "loss": 1.027, "step": 52300 }, { "epoch": 7.239568941696601, "grad_norm": 0.045472096651792526, "learning_rate": 8.406136172597366e-05, "loss": 1.0311, "step": 52400 }, { "epoch": 7.2533849129593815, "grad_norm": 0.0669686570763588, "learning_rate": 8.36410759316335e-05, "loss": 1.0309, "step": 52500 }, { "epoch": 7.267200884222161, "grad_norm": 0.0454520583152771, "learning_rate": 8.322079013729334e-05, "loss": 1.0327, "step": 52600 }, { "epoch": 7.281016855484941, "grad_norm": 0.05776028707623482, "learning_rate": 8.28005043429532e-05, "loss": 1.0318, "step": 52700 }, { "epoch": 7.2948328267477205, "grad_norm": 0.051905229687690735, "learning_rate": 8.238021854861305e-05, "loss": 1.0313, "step": 52800 }, { "epoch": 7.3086487980105, "grad_norm": 0.056912437081336975, "learning_rate": 8.19599327542729e-05, "loss": 1.0325, "step": 52900 }, { "epoch": 7.32246476927328, "grad_norm": 0.04940250515937805, "learning_rate": 8.153964695993274e-05, "loss": 1.0323, "step": 53000 }, { "epoch": 7.3362807405360595, "grad_norm": 0.04186444729566574, "learning_rate": 8.11193611655926e-05, "loss": 1.0285, "step": 53100 }, { "epoch": 7.350096711798839, "grad_norm": 0.041809357702732086, "learning_rate": 8.069907537125245e-05, "loss": 1.0289, "step": 53200 }, { "epoch": 7.363912683061619, "grad_norm": 0.05794375389814377, "learning_rate": 8.02787895769123e-05, "loss": 1.031, "step": 53300 }, { "epoch": 7.377728654324399, "grad_norm": 0.08333911001682281, "learning_rate": 7.985850378257213e-05, "loss": 1.0316, "step": 53400 }, { "epoch": 7.391544625587179, "grad_norm": 0.06473658233880997, "learning_rate": 7.943821798823199e-05, "loss": 1.0317, "step": 53500 }, { "epoch": 7.405360596849959, "grad_norm": 0.05173886939883232, "learning_rate": 7.901793219389184e-05, "loss": 1.0308, "step": 53600 }, { "epoch": 7.419176568112738, "grad_norm": 0.06362345069646835, "learning_rate": 7.859764639955169e-05, "loss": 1.0324, "step": 53700 }, { "epoch": 7.432992539375518, "grad_norm": 0.054053716361522675, "learning_rate": 7.817736060521152e-05, "loss": 1.0303, "step": 53800 }, { "epoch": 7.446808510638298, "grad_norm": 0.048420459032058716, "learning_rate": 7.775707481087139e-05, "loss": 1.0299, "step": 53900 }, { "epoch": 7.460624481901077, "grad_norm": 0.0606950968503952, "learning_rate": 7.733678901653123e-05, "loss": 1.0317, "step": 54000 }, { "epoch": 7.474440453163857, "grad_norm": 0.06072583049535751, "learning_rate": 7.691650322219108e-05, "loss": 1.033, "step": 54100 }, { "epoch": 7.488256424426638, "grad_norm": 0.05064817890524864, "learning_rate": 7.649621742785093e-05, "loss": 1.0287, "step": 54200 }, { "epoch": 7.502072395689417, "grad_norm": 0.09318757057189941, "learning_rate": 7.607593163351078e-05, "loss": 1.0296, "step": 54300 }, { "epoch": 7.515888366952197, "grad_norm": 0.0935215950012207, "learning_rate": 7.565564583917063e-05, "loss": 1.0322, "step": 54400 }, { "epoch": 7.529704338214977, "grad_norm": 0.07255256175994873, "learning_rate": 7.523536004483048e-05, "loss": 1.0333, "step": 54500 }, { "epoch": 7.543520309477756, "grad_norm": 0.05486008897423744, "learning_rate": 7.481507425049033e-05, "loss": 1.032, "step": 54600 }, { "epoch": 7.557336280740536, "grad_norm": 0.0525212287902832, "learning_rate": 7.439478845615017e-05, "loss": 1.0293, "step": 54700 }, { "epoch": 7.571152252003316, "grad_norm": 0.047569695860147476, "learning_rate": 7.397450266181002e-05, "loss": 1.0282, "step": 54800 }, { "epoch": 7.584968223266095, "grad_norm": 0.06165711581707001, "learning_rate": 7.355421686746987e-05, "loss": 1.0312, "step": 54900 }, { "epoch": 7.598784194528875, "grad_norm": 0.0578945092856884, "learning_rate": 7.313393107312972e-05, "loss": 1.0307, "step": 55000 }, { "epoch": 7.598784194528875, "eval_accuracy": 0.5305025000901846, "eval_loss": 1.0327985286712646, "eval_runtime": 731.5754, "eval_samples_per_second": 281.394, "eval_steps_per_second": 8.795, "step": 55000 }, { "epoch": 7.612600165791655, "grad_norm": 0.0795338973402977, "learning_rate": 7.271784813673297e-05, "loss": 1.0294, "step": 55100 }, { "epoch": 7.626416137054435, "grad_norm": 0.06103779003024101, "learning_rate": 7.229756234239283e-05, "loss": 1.033, "step": 55200 }, { "epoch": 7.640232108317215, "grad_norm": 0.0635315552353859, "learning_rate": 7.187727654805266e-05, "loss": 1.0296, "step": 55300 }, { "epoch": 7.654048079579995, "grad_norm": 0.05289231240749359, "learning_rate": 7.145699075371253e-05, "loss": 1.034, "step": 55400 }, { "epoch": 7.667864050842774, "grad_norm": 0.07801427692174911, "learning_rate": 7.103670495937236e-05, "loss": 1.0332, "step": 55500 }, { "epoch": 7.681680022105554, "grad_norm": 0.07564268261194229, "learning_rate": 7.061641916503222e-05, "loss": 1.0299, "step": 55600 }, { "epoch": 7.695495993368334, "grad_norm": 0.04168133810162544, "learning_rate": 7.019613337069206e-05, "loss": 1.03, "step": 55700 }, { "epoch": 7.709311964631113, "grad_norm": 0.11210035532712936, "learning_rate": 6.977584757635192e-05, "loss": 1.0301, "step": 55800 }, { "epoch": 7.723127935893894, "grad_norm": 0.09023060649633408, "learning_rate": 6.935556178201175e-05, "loss": 1.0285, "step": 55900 }, { "epoch": 7.7369439071566735, "grad_norm": 0.05271260067820549, "learning_rate": 6.893527598767162e-05, "loss": 1.0315, "step": 56000 }, { "epoch": 7.750759878419453, "grad_norm": 0.06293012201786041, "learning_rate": 6.851499019333145e-05, "loss": 1.0286, "step": 56100 }, { "epoch": 7.764575849682233, "grad_norm": 0.04555558040738106, "learning_rate": 6.809470439899131e-05, "loss": 1.0308, "step": 56200 }, { "epoch": 7.7783918209450125, "grad_norm": 0.042364273220300674, "learning_rate": 6.767441860465115e-05, "loss": 1.0311, "step": 56300 }, { "epoch": 7.792207792207792, "grad_norm": 0.05084213241934776, "learning_rate": 6.725413281031101e-05, "loss": 1.0298, "step": 56400 }, { "epoch": 7.806023763470572, "grad_norm": 0.059168051928281784, "learning_rate": 6.683384701597085e-05, "loss": 1.0303, "step": 56500 }, { "epoch": 7.8198397347333515, "grad_norm": 0.05535740405321121, "learning_rate": 6.641356122163071e-05, "loss": 1.0306, "step": 56600 }, { "epoch": 7.833655705996131, "grad_norm": 0.06625715643167496, "learning_rate": 6.599327542729054e-05, "loss": 1.0283, "step": 56700 }, { "epoch": 7.847471677258911, "grad_norm": 0.04644458368420601, "learning_rate": 6.55729896329504e-05, "loss": 1.0289, "step": 56800 }, { "epoch": 7.861287648521691, "grad_norm": 0.05319574847817421, "learning_rate": 6.515270383861024e-05, "loss": 1.0303, "step": 56900 }, { "epoch": 7.875103619784471, "grad_norm": 0.06394356489181519, "learning_rate": 6.47324180442701e-05, "loss": 1.0315, "step": 57000 }, { "epoch": 7.888919591047251, "grad_norm": 0.0535539835691452, "learning_rate": 6.431633510787335e-05, "loss": 1.0323, "step": 57100 }, { "epoch": 7.90273556231003, "grad_norm": 0.05220150947570801, "learning_rate": 6.38960493135332e-05, "loss": 1.032, "step": 57200 }, { "epoch": 7.91655153357281, "grad_norm": 0.04795517399907112, "learning_rate": 6.347576351919304e-05, "loss": 1.03, "step": 57300 }, { "epoch": 7.93036750483559, "grad_norm": 0.0748489499092102, "learning_rate": 6.30554777248529e-05, "loss": 1.0338, "step": 57400 }, { "epoch": 7.944183476098369, "grad_norm": 0.08164035528898239, "learning_rate": 6.263519193051274e-05, "loss": 1.0318, "step": 57500 }, { "epoch": 7.95799944736115, "grad_norm": 0.0764247477054596, "learning_rate": 6.221490613617259e-05, "loss": 1.0278, "step": 57600 }, { "epoch": 7.97181541862393, "grad_norm": 0.05609816685318947, "learning_rate": 6.179462034183244e-05, "loss": 1.0307, "step": 57700 }, { "epoch": 7.985631389886709, "grad_norm": 0.05001819133758545, "learning_rate": 6.137433454749229e-05, "loss": 1.0297, "step": 57800 }, { "epoch": 7.999447361149489, "grad_norm": 0.10084258019924164, "learning_rate": 6.0954048753152136e-05, "loss": 1.0339, "step": 57900 }, { "epoch": 8.013263332412269, "grad_norm": 0.07571733742952347, "learning_rate": 6.0533762958811985e-05, "loss": 1.0305, "step": 58000 }, { "epoch": 8.027079303675048, "grad_norm": 0.059294216334819794, "learning_rate": 6.011347716447183e-05, "loss": 1.026, "step": 58100 }, { "epoch": 8.040895274937828, "grad_norm": 0.04530787095427513, "learning_rate": 5.969319137013168e-05, "loss": 1.0282, "step": 58200 }, { "epoch": 8.054711246200608, "grad_norm": 0.05052864924073219, "learning_rate": 5.927290557579153e-05, "loss": 1.0271, "step": 58300 }, { "epoch": 8.068527217463387, "grad_norm": 0.04923342168331146, "learning_rate": 5.885261978145138e-05, "loss": 1.029, "step": 58400 }, { "epoch": 8.082343188726167, "grad_norm": 0.04905908182263374, "learning_rate": 5.843233398711123e-05, "loss": 1.0277, "step": 58500 }, { "epoch": 8.096159159988947, "grad_norm": 0.046151451766490936, "learning_rate": 5.801204819277108e-05, "loss": 1.0289, "step": 58600 }, { "epoch": 8.109975131251726, "grad_norm": 0.06011873856186867, "learning_rate": 5.7591762398430925e-05, "loss": 1.0245, "step": 58700 }, { "epoch": 8.123791102514506, "grad_norm": 0.06879663467407227, "learning_rate": 5.717147660409078e-05, "loss": 1.0271, "step": 58800 }, { "epoch": 8.137607073777286, "grad_norm": 0.04675479233264923, "learning_rate": 5.675119080975063e-05, "loss": 1.0263, "step": 58900 }, { "epoch": 8.151423045040067, "grad_norm": 0.08497285097837448, "learning_rate": 5.633090501541048e-05, "loss": 1.0287, "step": 59000 }, { "epoch": 8.165239016302847, "grad_norm": 0.07600156217813492, "learning_rate": 5.5910619221070326e-05, "loss": 1.0262, "step": 59100 }, { "epoch": 8.179054987565626, "grad_norm": 0.04951677843928337, "learning_rate": 5.549453628467357e-05, "loss": 1.0283, "step": 59200 }, { "epoch": 8.192870958828406, "grad_norm": 0.05662324279546738, "learning_rate": 5.507425049033342e-05, "loss": 1.0295, "step": 59300 }, { "epoch": 8.206686930091186, "grad_norm": 0.05791959911584854, "learning_rate": 5.465396469599327e-05, "loss": 1.0285, "step": 59400 }, { "epoch": 8.220502901353965, "grad_norm": 0.058768805116415024, "learning_rate": 5.423367890165312e-05, "loss": 1.0272, "step": 59500 }, { "epoch": 8.234318872616745, "grad_norm": 0.05399869754910469, "learning_rate": 5.381339310731297e-05, "loss": 1.0301, "step": 59600 }, { "epoch": 8.248134843879525, "grad_norm": 0.06434085965156555, "learning_rate": 5.3393107312972814e-05, "loss": 1.0277, "step": 59700 }, { "epoch": 8.261950815142304, "grad_norm": 0.054656483232975006, "learning_rate": 5.297282151863267e-05, "loss": 1.0295, "step": 59800 }, { "epoch": 8.275766786405084, "grad_norm": 0.04396641626954079, "learning_rate": 5.255253572429251e-05, "loss": 1.0276, "step": 59900 }, { "epoch": 8.289582757667864, "grad_norm": 0.058395449072122574, "learning_rate": 5.2132249929952366e-05, "loss": 1.0267, "step": 60000 }, { "epoch": 8.289582757667864, "eval_accuracy": 0.5312832658873073, "eval_loss": 1.0315501689910889, "eval_runtime": 729.415, "eval_samples_per_second": 282.228, "eval_steps_per_second": 8.821, "step": 60000 }, { "epoch": 8.303398728930643, "grad_norm": 0.06770013272762299, "learning_rate": 5.171196413561221e-05, "loss": 1.029, "step": 60100 }, { "epoch": 8.317214700193423, "grad_norm": 0.06161688268184662, "learning_rate": 5.1291678341272063e-05, "loss": 1.0242, "step": 60200 }, { "epoch": 8.331030671456203, "grad_norm": 0.04140911623835564, "learning_rate": 5.087139254693191e-05, "loss": 1.029, "step": 60300 }, { "epoch": 8.344846642718982, "grad_norm": 0.07091998308897018, "learning_rate": 5.045110675259176e-05, "loss": 1.0268, "step": 60400 }, { "epoch": 8.358662613981762, "grad_norm": 0.05135732889175415, "learning_rate": 5.003082095825161e-05, "loss": 1.0264, "step": 60500 }, { "epoch": 8.372478585244544, "grad_norm": 0.05828474089503288, "learning_rate": 4.961053516391146e-05, "loss": 1.0271, "step": 60600 }, { "epoch": 8.386294556507323, "grad_norm": 0.05920015275478363, "learning_rate": 4.9190249369571306e-05, "loss": 1.0263, "step": 60700 }, { "epoch": 8.400110527770103, "grad_norm": 0.048502273857593536, "learning_rate": 4.8769963575231155e-05, "loss": 1.029, "step": 60800 }, { "epoch": 8.413926499032883, "grad_norm": 0.049063604325056076, "learning_rate": 4.8349677780891e-05, "loss": 1.0294, "step": 60900 }, { "epoch": 8.427742470295662, "grad_norm": 0.05672093480825424, "learning_rate": 4.792939198655085e-05, "loss": 1.0297, "step": 61000 }, { "epoch": 8.441558441558442, "grad_norm": 0.06934633105993271, "learning_rate": 4.75091061922107e-05, "loss": 1.0261, "step": 61100 }, { "epoch": 8.455374412821222, "grad_norm": 0.04098910838365555, "learning_rate": 4.709302325581395e-05, "loss": 1.0292, "step": 61200 }, { "epoch": 8.469190384084001, "grad_norm": 0.06421385705471039, "learning_rate": 4.6672737461473794e-05, "loss": 1.0315, "step": 61300 }, { "epoch": 8.483006355346781, "grad_norm": 0.05238828435540199, "learning_rate": 4.625245166713365e-05, "loss": 1.0309, "step": 61400 }, { "epoch": 8.49682232660956, "grad_norm": 0.049910806119441986, "learning_rate": 4.583216587279349e-05, "loss": 1.0257, "step": 61500 }, { "epoch": 8.51063829787234, "grad_norm": 0.06672196090221405, "learning_rate": 4.541188007845335e-05, "loss": 1.0328, "step": 61600 }, { "epoch": 8.52445426913512, "grad_norm": 0.05466538295149803, "learning_rate": 4.4991594284113195e-05, "loss": 1.0284, "step": 61700 }, { "epoch": 8.5382702403979, "grad_norm": 0.05218784883618355, "learning_rate": 4.4571308489773044e-05, "loss": 1.0285, "step": 61800 }, { "epoch": 8.55208621166068, "grad_norm": 0.04263923689723015, "learning_rate": 4.415102269543289e-05, "loss": 1.0307, "step": 61900 }, { "epoch": 8.565902182923459, "grad_norm": 0.054478637874126434, "learning_rate": 4.373073690109274e-05, "loss": 1.0291, "step": 62000 }, { "epoch": 8.579718154186239, "grad_norm": 0.05667020007967949, "learning_rate": 4.331045110675259e-05, "loss": 1.0296, "step": 62100 }, { "epoch": 8.593534125449018, "grad_norm": 0.0490160770714283, "learning_rate": 4.289016531241244e-05, "loss": 1.029, "step": 62200 }, { "epoch": 8.607350096711798, "grad_norm": 0.049655403941869736, "learning_rate": 4.246987951807229e-05, "loss": 1.0298, "step": 62300 }, { "epoch": 8.62116606797458, "grad_norm": 0.047429408878088, "learning_rate": 4.2049593723732135e-05, "loss": 1.0277, "step": 62400 }, { "epoch": 8.634982039237359, "grad_norm": 0.05222218483686447, "learning_rate": 4.1629307929391984e-05, "loss": 1.0292, "step": 62500 }, { "epoch": 8.648798010500139, "grad_norm": 0.05841238424181938, "learning_rate": 4.120902213505183e-05, "loss": 1.029, "step": 62600 }, { "epoch": 8.662613981762918, "grad_norm": 0.0452195480465889, "learning_rate": 4.078873634071168e-05, "loss": 1.0265, "step": 62700 }, { "epoch": 8.676429953025698, "grad_norm": 0.049306340515613556, "learning_rate": 4.036845054637153e-05, "loss": 1.0308, "step": 62800 }, { "epoch": 8.690245924288478, "grad_norm": 0.050401389598846436, "learning_rate": 3.994816475203138e-05, "loss": 1.0294, "step": 62900 }, { "epoch": 8.704061895551257, "grad_norm": 0.04503024369478226, "learning_rate": 3.952787895769123e-05, "loss": 1.0291, "step": 63000 }, { "epoch": 8.717877866814037, "grad_norm": 0.0738733783364296, "learning_rate": 3.9107593163351075e-05, "loss": 1.0279, "step": 63100 }, { "epoch": 8.731693838076817, "grad_norm": 0.04586975276470184, "learning_rate": 3.869151022695433e-05, "loss": 1.026, "step": 63200 }, { "epoch": 8.745509809339596, "grad_norm": 0.04988343268632889, "learning_rate": 3.8271224432614176e-05, "loss": 1.0257, "step": 63300 }, { "epoch": 8.759325780602376, "grad_norm": 0.07822008430957794, "learning_rate": 3.7850938638274025e-05, "loss": 1.0254, "step": 63400 }, { "epoch": 8.773141751865156, "grad_norm": 0.058496229350566864, "learning_rate": 3.743065284393387e-05, "loss": 1.0263, "step": 63500 }, { "epoch": 8.786957723127935, "grad_norm": 0.04458677023649216, "learning_rate": 3.701036704959372e-05, "loss": 1.0292, "step": 63600 }, { "epoch": 8.800773694390715, "grad_norm": 0.06616061180830002, "learning_rate": 3.659008125525357e-05, "loss": 1.0309, "step": 63700 }, { "epoch": 8.814589665653495, "grad_norm": 0.06473194807767868, "learning_rate": 3.616979546091342e-05, "loss": 1.0265, "step": 63800 }, { "epoch": 8.828405636916274, "grad_norm": 0.047700874507427216, "learning_rate": 3.574950966657327e-05, "loss": 1.0303, "step": 63900 }, { "epoch": 8.842221608179056, "grad_norm": 0.055733323097229004, "learning_rate": 3.5329223872233116e-05, "loss": 1.0279, "step": 64000 }, { "epoch": 8.856037579441836, "grad_norm": 0.04398791491985321, "learning_rate": 3.4908938077892965e-05, "loss": 1.0284, "step": 64100 }, { "epoch": 8.869853550704615, "grad_norm": 0.08901511132717133, "learning_rate": 3.448865228355281e-05, "loss": 1.0283, "step": 64200 }, { "epoch": 8.883669521967395, "grad_norm": 0.05853118374943733, "learning_rate": 3.406836648921266e-05, "loss": 1.0291, "step": 64300 }, { "epoch": 8.897485493230175, "grad_norm": 0.043922308832407, "learning_rate": 3.364808069487251e-05, "loss": 1.0294, "step": 64400 }, { "epoch": 8.911301464492954, "grad_norm": 0.04332153871655464, "learning_rate": 3.322779490053236e-05, "loss": 1.0277, "step": 64500 }, { "epoch": 8.925117435755734, "grad_norm": 0.09197825193405151, "learning_rate": 3.280750910619221e-05, "loss": 1.0295, "step": 64600 }, { "epoch": 8.938933407018514, "grad_norm": 0.05589272826910019, "learning_rate": 3.2387223311852056e-05, "loss": 1.0274, "step": 64700 }, { "epoch": 8.952749378281293, "grad_norm": 0.06028933823108673, "learning_rate": 3.1966937517511904e-05, "loss": 1.0285, "step": 64800 }, { "epoch": 8.966565349544073, "grad_norm": 0.05357721447944641, "learning_rate": 3.154665172317175e-05, "loss": 1.027, "step": 64900 }, { "epoch": 8.980381320806853, "grad_norm": 0.07362578809261322, "learning_rate": 3.11263659288316e-05, "loss": 1.0273, "step": 65000 }, { "epoch": 8.980381320806853, "eval_accuracy": 0.5319501927585898, "eval_loss": 1.0305662155151367, "eval_runtime": 722.9505, "eval_samples_per_second": 284.751, "eval_steps_per_second": 8.9, "step": 65000 }, { "epoch": 8.994197292069632, "grad_norm": 0.04831722378730774, "learning_rate": 3.070608013449145e-05, "loss": 1.0294, "step": 65100 }, { "epoch": 9.008013263332412, "grad_norm": 0.06001870334148407, "learning_rate": 3.0289997198094702e-05, "loss": 1.0306, "step": 65200 }, { "epoch": 9.021829234595192, "grad_norm": 0.04466562718153, "learning_rate": 2.986971140375455e-05, "loss": 1.0267, "step": 65300 }, { "epoch": 9.035645205857971, "grad_norm": 0.059990085661411285, "learning_rate": 2.94494256094144e-05, "loss": 1.0248, "step": 65400 }, { "epoch": 9.049461177120751, "grad_norm": 0.05244195833802223, "learning_rate": 2.9029139815074248e-05, "loss": 1.0282, "step": 65500 }, { "epoch": 9.06327714838353, "grad_norm": 0.060148317366838455, "learning_rate": 2.8608854020734097e-05, "loss": 1.0266, "step": 65600 }, { "epoch": 9.07709311964631, "grad_norm": 0.051530059427022934, "learning_rate": 2.8188568226393945e-05, "loss": 1.0257, "step": 65700 }, { "epoch": 9.090909090909092, "grad_norm": 0.06650034338235855, "learning_rate": 2.7768282432053794e-05, "loss": 1.0276, "step": 65800 }, { "epoch": 9.104725062171871, "grad_norm": 0.04850700497627258, "learning_rate": 2.7347996637713642e-05, "loss": 1.0249, "step": 65900 }, { "epoch": 9.118541033434651, "grad_norm": 0.057128727436065674, "learning_rate": 2.692771084337349e-05, "loss": 1.0264, "step": 66000 }, { "epoch": 9.13235700469743, "grad_norm": 0.056875213980674744, "learning_rate": 2.650742504903334e-05, "loss": 1.0285, "step": 66100 }, { "epoch": 9.14617297596021, "grad_norm": 0.05632421374320984, "learning_rate": 2.6087139254693188e-05, "loss": 1.0286, "step": 66200 }, { "epoch": 9.15998894722299, "grad_norm": 0.04903789609670639, "learning_rate": 2.5666853460353037e-05, "loss": 1.0233, "step": 66300 }, { "epoch": 9.17380491848577, "grad_norm": 0.04932420328259468, "learning_rate": 2.5246567666012885e-05, "loss": 1.0273, "step": 66400 }, { "epoch": 9.18762088974855, "grad_norm": 0.0668862909078598, "learning_rate": 2.4826281871672734e-05, "loss": 1.0264, "step": 66500 }, { "epoch": 9.201436861011329, "grad_norm": 0.05283021926879883, "learning_rate": 2.4405996077332586e-05, "loss": 1.0278, "step": 66600 }, { "epoch": 9.215252832274109, "grad_norm": 0.04914732649922371, "learning_rate": 2.3985710282992434e-05, "loss": 1.0276, "step": 66700 }, { "epoch": 9.229068803536888, "grad_norm": 0.06511181592941284, "learning_rate": 2.3565424488652283e-05, "loss": 1.0268, "step": 66800 }, { "epoch": 9.242884774799668, "grad_norm": 0.06101306900382042, "learning_rate": 2.314513869431213e-05, "loss": 1.0267, "step": 66900 }, { "epoch": 9.256700746062448, "grad_norm": 0.05272289365530014, "learning_rate": 2.272485289997198e-05, "loss": 1.0242, "step": 67000 }, { "epoch": 9.270516717325227, "grad_norm": 0.04828105494379997, "learning_rate": 2.230456710563183e-05, "loss": 1.0258, "step": 67100 }, { "epoch": 9.284332688588007, "grad_norm": 0.054294098168611526, "learning_rate": 2.1888484169235077e-05, "loss": 1.0262, "step": 67200 }, { "epoch": 9.298148659850787, "grad_norm": 0.04951765388250351, "learning_rate": 2.1468198374894926e-05, "loss": 1.0254, "step": 67300 }, { "epoch": 9.311964631113566, "grad_norm": 0.047647446393966675, "learning_rate": 2.1047912580554774e-05, "loss": 1.0262, "step": 67400 }, { "epoch": 9.325780602376348, "grad_norm": 0.062047079205513, "learning_rate": 2.0627626786214623e-05, "loss": 1.0287, "step": 67500 }, { "epoch": 9.339596573639128, "grad_norm": 0.05751033127307892, "learning_rate": 2.020734099187447e-05, "loss": 1.027, "step": 67600 }, { "epoch": 9.353412544901907, "grad_norm": 0.058642346411943436, "learning_rate": 1.978705519753432e-05, "loss": 1.0276, "step": 67700 }, { "epoch": 9.367228516164687, "grad_norm": 0.050882838666439056, "learning_rate": 1.936676940319417e-05, "loss": 1.0223, "step": 67800 }, { "epoch": 9.381044487427467, "grad_norm": 0.053814638406038284, "learning_rate": 1.8946483608854017e-05, "loss": 1.0271, "step": 67900 }, { "epoch": 9.394860458690246, "grad_norm": 0.05407179519534111, "learning_rate": 1.852619781451387e-05, "loss": 1.0242, "step": 68000 }, { "epoch": 9.408676429953026, "grad_norm": 0.05431421846151352, "learning_rate": 1.8105912020173718e-05, "loss": 1.0246, "step": 68100 }, { "epoch": 9.422492401215806, "grad_norm": 0.05826635658740997, "learning_rate": 1.7685626225833566e-05, "loss": 1.024, "step": 68200 }, { "epoch": 9.436308372478585, "grad_norm": 0.043603766709566116, "learning_rate": 1.7265340431493415e-05, "loss": 1.025, "step": 68300 }, { "epoch": 9.450124343741365, "grad_norm": 0.0555894561111927, "learning_rate": 1.6845054637153263e-05, "loss": 1.0267, "step": 68400 }, { "epoch": 9.463940315004145, "grad_norm": 0.046029891818761826, "learning_rate": 1.6424768842813112e-05, "loss": 1.0247, "step": 68500 }, { "epoch": 9.477756286266924, "grad_norm": 0.04906938225030899, "learning_rate": 1.600448304847296e-05, "loss": 1.0233, "step": 68600 }, { "epoch": 9.491572257529704, "grad_norm": 0.07827210426330566, "learning_rate": 1.558419725413281e-05, "loss": 1.0262, "step": 68700 }, { "epoch": 9.505388228792484, "grad_norm": 0.04391390085220337, "learning_rate": 1.5163911459792658e-05, "loss": 1.0255, "step": 68800 }, { "epoch": 9.519204200055263, "grad_norm": 0.05310402810573578, "learning_rate": 1.4743625665452506e-05, "loss": 1.0268, "step": 68900 }, { "epoch": 9.533020171318043, "grad_norm": 0.060242168605327606, "learning_rate": 1.4323339871112355e-05, "loss": 1.0257, "step": 69000 }, { "epoch": 9.546836142580823, "grad_norm": 0.04949665814638138, "learning_rate": 1.3903054076772205e-05, "loss": 1.0294, "step": 69100 }, { "epoch": 9.560652113843604, "grad_norm": 0.05413687229156494, "learning_rate": 1.3482768282432054e-05, "loss": 1.0272, "step": 69200 }, { "epoch": 9.574468085106384, "grad_norm": 0.05380227789282799, "learning_rate": 1.3062482488091902e-05, "loss": 1.025, "step": 69300 }, { "epoch": 9.588284056369163, "grad_norm": 0.04961249604821205, "learning_rate": 1.2646399551695151e-05, "loss": 1.0289, "step": 69400 }, { "epoch": 9.602100027631943, "grad_norm": 0.045629873871803284, "learning_rate": 1.2226113757355e-05, "loss": 1.0269, "step": 69500 }, { "epoch": 9.615915998894723, "grad_norm": 0.04661751165986061, "learning_rate": 1.1805827963014848e-05, "loss": 1.0277, "step": 69600 }, { "epoch": 9.629731970157502, "grad_norm": 0.06289409101009369, "learning_rate": 1.1385542168674697e-05, "loss": 1.0246, "step": 69700 }, { "epoch": 9.643547941420282, "grad_norm": 0.061526406556367874, "learning_rate": 1.0965256374334547e-05, "loss": 1.0252, "step": 69800 }, { "epoch": 9.657363912683062, "grad_norm": 0.05611636862158775, "learning_rate": 1.0544970579994395e-05, "loss": 1.0281, "step": 69900 }, { "epoch": 9.671179883945841, "grad_norm": 0.05305150896310806, "learning_rate": 1.0124684785654244e-05, "loss": 1.027, "step": 70000 }, { "epoch": 9.671179883945841, "eval_accuracy": 0.5323623139821072, "eval_loss": 1.029943823814392, "eval_runtime": 726.2479, "eval_samples_per_second": 283.458, "eval_steps_per_second": 8.859, "step": 70000 }, { "epoch": 9.684995855208621, "grad_norm": 0.06483161449432373, "learning_rate": 9.704398991314093e-06, "loss": 1.0262, "step": 70100 }, { "epoch": 9.6988118264714, "grad_norm": 0.05063271522521973, "learning_rate": 9.284113196973941e-06, "loss": 1.0246, "step": 70200 }, { "epoch": 9.71262779773418, "grad_norm": 0.04985768347978592, "learning_rate": 8.86382740263379e-06, "loss": 1.03, "step": 70300 }, { "epoch": 9.72644376899696, "grad_norm": 0.04751725122332573, "learning_rate": 8.443541608293638e-06, "loss": 1.0272, "step": 70400 }, { "epoch": 9.74025974025974, "grad_norm": 0.042586106806993484, "learning_rate": 8.023255813953487e-06, "loss": 1.0254, "step": 70500 }, { "epoch": 9.75407571152252, "grad_norm": 0.059688206762075424, "learning_rate": 7.602970019613336e-06, "loss": 1.0255, "step": 70600 }, { "epoch": 9.767891682785299, "grad_norm": 0.04823042452335358, "learning_rate": 7.182684225273185e-06, "loss": 1.028, "step": 70700 }, { "epoch": 9.78170765404808, "grad_norm": 0.0480177104473114, "learning_rate": 6.762398430933033e-06, "loss": 1.025, "step": 70800 }, { "epoch": 9.79552362531086, "grad_norm": 0.045797545462846756, "learning_rate": 6.342112636592882e-06, "loss": 1.023, "step": 70900 }, { "epoch": 9.80933959657364, "grad_norm": 0.04858710244297981, "learning_rate": 5.921826842252732e-06, "loss": 1.0216, "step": 71000 }, { "epoch": 9.82315556783642, "grad_norm": 0.05248698219656944, "learning_rate": 5.501541047912581e-06, "loss": 1.0252, "step": 71100 }, { "epoch": 9.8369715390992, "grad_norm": 0.045856546610593796, "learning_rate": 5.081255253572429e-06, "loss": 1.0249, "step": 71200 }, { "epoch": 9.850787510361979, "grad_norm": 0.047852564603090286, "learning_rate": 4.660969459232278e-06, "loss": 1.0238, "step": 71300 }, { "epoch": 9.864603481624759, "grad_norm": 0.044457610696554184, "learning_rate": 4.240683664892126e-06, "loss": 1.0245, "step": 71400 }, { "epoch": 9.878419452887538, "grad_norm": 0.06768154352903366, "learning_rate": 3.824600728495377e-06, "loss": 1.0256, "step": 71500 }, { "epoch": 9.892235424150318, "grad_norm": 0.050749246031045914, "learning_rate": 3.4043149341552255e-06, "loss": 1.0224, "step": 71600 }, { "epoch": 9.906051395413098, "grad_norm": 0.04643206670880318, "learning_rate": 2.984029139815074e-06, "loss": 1.0279, "step": 71700 }, { "epoch": 9.919867366675877, "grad_norm": 0.05305636674165726, "learning_rate": 2.5637433454749226e-06, "loss": 1.0257, "step": 71800 }, { "epoch": 9.933683337938657, "grad_norm": 0.06888972967863083, "learning_rate": 2.1434575511347716e-06, "loss": 1.0301, "step": 71900 }, { "epoch": 9.947499309201437, "grad_norm": 0.04633474349975586, "learning_rate": 1.7231717567946201e-06, "loss": 1.0228, "step": 72000 }, { "epoch": 9.961315280464216, "grad_norm": 0.05391710251569748, "learning_rate": 1.302885962454469e-06, "loss": 1.0284, "step": 72100 }, { "epoch": 9.975131251726996, "grad_norm": 0.048064954578876495, "learning_rate": 8.826001681143176e-07, "loss": 1.0257, "step": 72200 }, { "epoch": 9.988947222989776, "grad_norm": 0.04865507408976555, "learning_rate": 4.623143737741664e-07, "loss": 1.0274, "step": 72300 } ], "logging_steps": 100, "max_steps": 72380, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 9.683106445125485e+18, "train_batch_size": 64, "trial_name": null, "trial_params": null }