{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.537318712415989, "eval_steps": 1000, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.017686593562079942, "grad_norm": 4.990772724151611, "learning_rate": 5.000000000000001e-07, "loss": 0.9549, "step": 25 }, { "epoch": 0.035373187124159884, "grad_norm": 4.042325019836426, "learning_rate": 1.0000000000000002e-06, "loss": 0.8294, "step": 50 }, { "epoch": 0.05305978068623983, "grad_norm": 3.429729461669922, "learning_rate": 1.5e-06, "loss": 0.5478, "step": 75 }, { "epoch": 0.07074637424831977, "grad_norm": 3.084613800048828, "learning_rate": 2.0000000000000003e-06, "loss": 0.482, "step": 100 }, { "epoch": 0.08843296781039972, "grad_norm": 3.2256970405578613, "learning_rate": 2.5e-06, "loss": 0.4206, "step": 125 }, { "epoch": 0.10611956137247966, "grad_norm": 3.0460891723632812, "learning_rate": 3e-06, "loss": 0.4171, "step": 150 }, { "epoch": 0.1238061549345596, "grad_norm": 2.976763963699341, "learning_rate": 3.5e-06, "loss": 0.3896, "step": 175 }, { "epoch": 0.14149274849663954, "grad_norm": 2.8915836811065674, "learning_rate": 4.000000000000001e-06, "loss": 0.4149, "step": 200 }, { "epoch": 0.1591793420587195, "grad_norm": 3.01548433303833, "learning_rate": 4.5e-06, "loss": 0.3811, "step": 225 }, { "epoch": 0.17686593562079944, "grad_norm": 3.8455846309661865, "learning_rate": 5e-06, "loss": 0.3872, "step": 250 }, { "epoch": 0.19455252918287938, "grad_norm": 2.805778741836548, "learning_rate": 5.500000000000001e-06, "loss": 0.3714, "step": 275 }, { "epoch": 0.21223912274495932, "grad_norm": 3.138227939605713, "learning_rate": 6e-06, "loss": 0.3768, "step": 300 }, { "epoch": 0.22992571630703926, "grad_norm": 3.5028634071350098, "learning_rate": 6.5000000000000004e-06, "loss": 0.3835, "step": 325 }, { "epoch": 0.2476123098691192, "grad_norm": 3.8886911869049072, "learning_rate": 7e-06, "loss": 0.359, "step": 350 }, { "epoch": 0.26529890343119916, "grad_norm": 2.8521852493286133, "learning_rate": 7.500000000000001e-06, "loss": 0.3507, "step": 375 }, { "epoch": 0.2829854969932791, "grad_norm": 2.901258945465088, "learning_rate": 8.000000000000001e-06, "loss": 0.3586, "step": 400 }, { "epoch": 0.30067209055535904, "grad_norm": 2.550757884979248, "learning_rate": 8.5e-06, "loss": 0.3658, "step": 425 }, { "epoch": 0.318358684117439, "grad_norm": 3.0283052921295166, "learning_rate": 9e-06, "loss": 0.3527, "step": 450 }, { "epoch": 0.3360452776795189, "grad_norm": 3.272336721420288, "learning_rate": 9.5e-06, "loss": 0.3674, "step": 475 }, { "epoch": 0.3537318712415989, "grad_norm": 2.8109354972839355, "learning_rate": 1e-05, "loss": 0.3592, "step": 500 }, { "epoch": 0.3714184648036788, "grad_norm": 2.21569561958313, "learning_rate": 9.944444444444445e-06, "loss": 0.3219, "step": 525 }, { "epoch": 0.38910505836575876, "grad_norm": 2.5520081520080566, "learning_rate": 9.88888888888889e-06, "loss": 0.3681, "step": 550 }, { "epoch": 0.40679165192783867, "grad_norm": 2.903778076171875, "learning_rate": 9.833333333333333e-06, "loss": 0.3424, "step": 575 }, { "epoch": 0.42447824548991864, "grad_norm": 2.4832980632781982, "learning_rate": 9.777777777777779e-06, "loss": 0.3269, "step": 600 }, { "epoch": 0.4421648390519986, "grad_norm": 2.759129285812378, "learning_rate": 9.722222222222223e-06, "loss": 0.327, "step": 625 }, { "epoch": 0.4598514326140785, "grad_norm": 2.3846707344055176, "learning_rate": 9.666666666666667e-06, "loss": 0.3367, "step": 650 }, { "epoch": 0.4775380261761585, "grad_norm": 2.462186574935913, "learning_rate": 9.611111111111112e-06, "loss": 0.3175, "step": 675 }, { "epoch": 0.4952246197382384, "grad_norm": 2.5651514530181885, "learning_rate": 9.555555555555556e-06, "loss": 0.3156, "step": 700 }, { "epoch": 0.5129112133003184, "grad_norm": 2.6540584564208984, "learning_rate": 9.5e-06, "loss": 0.3098, "step": 725 }, { "epoch": 0.5305978068623983, "grad_norm": 2.713265895843506, "learning_rate": 9.444444444444445e-06, "loss": 0.3226, "step": 750 }, { "epoch": 0.5482844004244782, "grad_norm": 2.128851890563965, "learning_rate": 9.38888888888889e-06, "loss": 0.3244, "step": 775 }, { "epoch": 0.5659709939865581, "grad_norm": 2.5764191150665283, "learning_rate": 9.333333333333334e-06, "loss": 0.3156, "step": 800 }, { "epoch": 0.5836575875486382, "grad_norm": 2.1402037143707275, "learning_rate": 9.277777777777778e-06, "loss": 0.3029, "step": 825 }, { "epoch": 0.6013441811107181, "grad_norm": 2.4773480892181396, "learning_rate": 9.222222222222224e-06, "loss": 0.293, "step": 850 }, { "epoch": 0.619030774672798, "grad_norm": 2.6644790172576904, "learning_rate": 9.166666666666666e-06, "loss": 0.3111, "step": 875 }, { "epoch": 0.636717368234878, "grad_norm": 2.3241961002349854, "learning_rate": 9.111111111111112e-06, "loss": 0.2726, "step": 900 }, { "epoch": 0.6544039617969579, "grad_norm": 2.6867661476135254, "learning_rate": 9.055555555555556e-06, "loss": 0.3083, "step": 925 }, { "epoch": 0.6720905553590378, "grad_norm": 2.138195037841797, "learning_rate": 9e-06, "loss": 0.3019, "step": 950 }, { "epoch": 0.6897771489211177, "grad_norm": 2.6360018253326416, "learning_rate": 8.944444444444446e-06, "loss": 0.2917, "step": 975 }, { "epoch": 0.7074637424831978, "grad_norm": 2.5191752910614014, "learning_rate": 8.888888888888888e-06, "loss": 0.294, "step": 1000 }, { "epoch": 0.7074637424831978, "eval_loss": 0.30541032552719116, "eval_runtime": 4620.7401, "eval_samples_per_second": 2.329, "eval_steps_per_second": 0.146, "eval_wer": 0.21139403895799108, "step": 1000 }, { "epoch": 0.7251503360452777, "grad_norm": 2.696348190307617, "learning_rate": 8.833333333333334e-06, "loss": 0.2912, "step": 1025 }, { "epoch": 0.7428369296073576, "grad_norm": 2.3700759410858154, "learning_rate": 8.777777777777778e-06, "loss": 0.2756, "step": 1050 }, { "epoch": 0.7605235231694376, "grad_norm": 2.314389705657959, "learning_rate": 8.722222222222224e-06, "loss": 0.2881, "step": 1075 }, { "epoch": 0.7782101167315175, "grad_norm": 2.3540844917297363, "learning_rate": 8.666666666666668e-06, "loss": 0.291, "step": 1100 }, { "epoch": 0.7958967102935974, "grad_norm": 2.3578970432281494, "learning_rate": 8.611111111111112e-06, "loss": 0.2846, "step": 1125 }, { "epoch": 0.8135833038556773, "grad_norm": 1.976126790046692, "learning_rate": 8.555555555555556e-06, "loss": 0.27, "step": 1150 }, { "epoch": 0.8312698974177574, "grad_norm": 2.194822072982788, "learning_rate": 8.5e-06, "loss": 0.2763, "step": 1175 }, { "epoch": 0.8489564909798373, "grad_norm": 2.1482298374176025, "learning_rate": 8.444444444444446e-06, "loss": 0.2731, "step": 1200 }, { "epoch": 0.8666430845419172, "grad_norm": 2.163198471069336, "learning_rate": 8.38888888888889e-06, "loss": 0.2602, "step": 1225 }, { "epoch": 0.8843296781039972, "grad_norm": 2.482328414916992, "learning_rate": 8.333333333333334e-06, "loss": 0.2896, "step": 1250 }, { "epoch": 0.9020162716660771, "grad_norm": 2.2900679111480713, "learning_rate": 8.277777777777778e-06, "loss": 0.2472, "step": 1275 }, { "epoch": 0.919702865228157, "grad_norm": 2.4348435401916504, "learning_rate": 8.222222222222222e-06, "loss": 0.2754, "step": 1300 }, { "epoch": 0.937389458790237, "grad_norm": 2.4213294982910156, "learning_rate": 8.166666666666668e-06, "loss": 0.2585, "step": 1325 }, { "epoch": 0.955076052352317, "grad_norm": 2.6255362033843994, "learning_rate": 8.111111111111112e-06, "loss": 0.25, "step": 1350 }, { "epoch": 0.9727626459143969, "grad_norm": 2.187410354614258, "learning_rate": 8.055555555555557e-06, "loss": 0.2809, "step": 1375 }, { "epoch": 0.9904492394764768, "grad_norm": 2.669609546661377, "learning_rate": 8.000000000000001e-06, "loss": 0.2707, "step": 1400 }, { "epoch": 1.0081358330385568, "grad_norm": 1.841189980506897, "learning_rate": 7.944444444444445e-06, "loss": 0.2272, "step": 1425 }, { "epoch": 1.0258224266006368, "grad_norm": 1.8432203531265259, "learning_rate": 7.88888888888889e-06, "loss": 0.2036, "step": 1450 }, { "epoch": 1.0435090201627166, "grad_norm": 1.961900234222412, "learning_rate": 7.833333333333333e-06, "loss": 0.1758, "step": 1475 }, { "epoch": 1.0611956137247966, "grad_norm": 1.7587320804595947, "learning_rate": 7.77777777777778e-06, "loss": 0.1847, "step": 1500 }, { "epoch": 1.0788822072868764, "grad_norm": 2.330167531967163, "learning_rate": 7.722222222222223e-06, "loss": 0.1602, "step": 1525 }, { "epoch": 1.0965688008489565, "grad_norm": 2.3043272495269775, "learning_rate": 7.666666666666667e-06, "loss": 0.1772, "step": 1550 }, { "epoch": 1.1142553944110365, "grad_norm": 2.093520164489746, "learning_rate": 7.611111111111111e-06, "loss": 0.1742, "step": 1575 }, { "epoch": 1.1319419879731163, "grad_norm": 2.2821807861328125, "learning_rate": 7.555555555555556e-06, "loss": 0.1685, "step": 1600 }, { "epoch": 1.1496285815351963, "grad_norm": 2.087881326675415, "learning_rate": 7.500000000000001e-06, "loss": 0.1722, "step": 1625 }, { "epoch": 1.1673151750972763, "grad_norm": 1.7071295976638794, "learning_rate": 7.444444444444445e-06, "loss": 0.168, "step": 1650 }, { "epoch": 1.1850017686593561, "grad_norm": 2.3514926433563232, "learning_rate": 7.38888888888889e-06, "loss": 0.1786, "step": 1675 }, { "epoch": 1.2026883622214362, "grad_norm": 2.0674052238464355, "learning_rate": 7.333333333333333e-06, "loss": 0.1625, "step": 1700 }, { "epoch": 1.2203749557835162, "grad_norm": 1.9389206171035767, "learning_rate": 7.277777777777778e-06, "loss": 0.1686, "step": 1725 }, { "epoch": 1.238061549345596, "grad_norm": 2.185607433319092, "learning_rate": 7.222222222222223e-06, "loss": 0.168, "step": 1750 }, { "epoch": 1.255748142907676, "grad_norm": 1.953815221786499, "learning_rate": 7.166666666666667e-06, "loss": 0.1634, "step": 1775 }, { "epoch": 1.2734347364697558, "grad_norm": 2.0842204093933105, "learning_rate": 7.111111111111112e-06, "loss": 0.1749, "step": 1800 }, { "epoch": 1.2911213300318358, "grad_norm": 2.197368621826172, "learning_rate": 7.055555555555557e-06, "loss": 0.173, "step": 1825 }, { "epoch": 1.3088079235939158, "grad_norm": 1.8195652961730957, "learning_rate": 7e-06, "loss": 0.1575, "step": 1850 }, { "epoch": 1.3264945171559956, "grad_norm": 1.9379323720932007, "learning_rate": 6.944444444444445e-06, "loss": 0.1591, "step": 1875 }, { "epoch": 1.3441811107180757, "grad_norm": 1.8081250190734863, "learning_rate": 6.88888888888889e-06, "loss": 0.1534, "step": 1900 }, { "epoch": 1.3618677042801557, "grad_norm": 2.040506601333618, "learning_rate": 6.833333333333334e-06, "loss": 0.1548, "step": 1925 }, { "epoch": 1.3795542978422355, "grad_norm": 1.8651851415634155, "learning_rate": 6.777777777777779e-06, "loss": 0.161, "step": 1950 }, { "epoch": 1.3972408914043155, "grad_norm": 1.9047472476959229, "learning_rate": 6.7222222222222235e-06, "loss": 0.1651, "step": 1975 }, { "epoch": 1.4149274849663955, "grad_norm": 1.6781543493270874, "learning_rate": 6.666666666666667e-06, "loss": 0.1572, "step": 2000 }, { "epoch": 1.4149274849663955, "eval_loss": 0.27680015563964844, "eval_runtime": 4951.943, "eval_samples_per_second": 2.173, "eval_steps_per_second": 0.136, "eval_wer": 0.18979308456543847, "step": 2000 }, { "epoch": 1.4326140785284753, "grad_norm": 1.9535053968429565, "learning_rate": 6.6111111111111115e-06, "loss": 0.1586, "step": 2025 }, { "epoch": 1.4503006720905554, "grad_norm": 1.7220150232315063, "learning_rate": 6.555555555555556e-06, "loss": 0.1634, "step": 2050 }, { "epoch": 1.4679872656526354, "grad_norm": 2.427858352661133, "learning_rate": 6.5000000000000004e-06, "loss": 0.1659, "step": 2075 }, { "epoch": 1.4856738592147152, "grad_norm": 1.9716796875, "learning_rate": 6.444444444444445e-06, "loss": 0.1511, "step": 2100 }, { "epoch": 1.5033604527767952, "grad_norm": 1.8595249652862549, "learning_rate": 6.3888888888888885e-06, "loss": 0.1556, "step": 2125 }, { "epoch": 1.5210470463388752, "grad_norm": 2.3424742221832275, "learning_rate": 6.333333333333333e-06, "loss": 0.1717, "step": 2150 }, { "epoch": 1.538733639900955, "grad_norm": 2.172686815261841, "learning_rate": 6.277777777777778e-06, "loss": 0.1586, "step": 2175 }, { "epoch": 1.556420233463035, "grad_norm": 1.7644214630126953, "learning_rate": 6.222222222222223e-06, "loss": 0.1521, "step": 2200 }, { "epoch": 1.574106827025115, "grad_norm": 2.1562883853912354, "learning_rate": 6.166666666666667e-06, "loss": 0.1578, "step": 2225 }, { "epoch": 1.5917934205871949, "grad_norm": 2.087270736694336, "learning_rate": 6.111111111111112e-06, "loss": 0.1556, "step": 2250 }, { "epoch": 1.6094800141492749, "grad_norm": 2.0405049324035645, "learning_rate": 6.055555555555555e-06, "loss": 0.1654, "step": 2275 }, { "epoch": 1.627166607711355, "grad_norm": 1.6984881162643433, "learning_rate": 6e-06, "loss": 0.1573, "step": 2300 }, { "epoch": 1.6448532012734347, "grad_norm": 2.290949583053589, "learning_rate": 5.944444444444445e-06, "loss": 0.1655, "step": 2325 }, { "epoch": 1.6625397948355147, "grad_norm": 1.9502941370010376, "learning_rate": 5.88888888888889e-06, "loss": 0.1462, "step": 2350 }, { "epoch": 1.6802263883975948, "grad_norm": 1.9741941690444946, "learning_rate": 5.833333333333334e-06, "loss": 0.1636, "step": 2375 }, { "epoch": 1.6979129819596745, "grad_norm": 1.9639830589294434, "learning_rate": 5.777777777777778e-06, "loss": 0.155, "step": 2400 }, { "epoch": 1.7155995755217543, "grad_norm": 1.6349612474441528, "learning_rate": 5.722222222222222e-06, "loss": 0.1659, "step": 2425 }, { "epoch": 1.7332861690838346, "grad_norm": 1.8674352169036865, "learning_rate": 5.666666666666667e-06, "loss": 0.152, "step": 2450 }, { "epoch": 1.7509727626459144, "grad_norm": 2.1415135860443115, "learning_rate": 5.611111111111112e-06, "loss": 0.1699, "step": 2475 }, { "epoch": 1.7686593562079942, "grad_norm": 1.5318387746810913, "learning_rate": 5.555555555555557e-06, "loss": 0.1592, "step": 2500 }, { "epoch": 1.7863459497700744, "grad_norm": 1.758480191230774, "learning_rate": 5.500000000000001e-06, "loss": 0.1537, "step": 2525 }, { "epoch": 1.8040325433321542, "grad_norm": 1.815434455871582, "learning_rate": 5.444444444444445e-06, "loss": 0.1533, "step": 2550 }, { "epoch": 1.821719136894234, "grad_norm": 1.5665532350540161, "learning_rate": 5.388888888888889e-06, "loss": 0.1601, "step": 2575 }, { "epoch": 1.839405730456314, "grad_norm": 2.0574448108673096, "learning_rate": 5.333333333333334e-06, "loss": 0.1548, "step": 2600 }, { "epoch": 1.857092324018394, "grad_norm": 1.7183318138122559, "learning_rate": 5.2777777777777785e-06, "loss": 0.1477, "step": 2625 }, { "epoch": 1.8747789175804739, "grad_norm": 1.7662074565887451, "learning_rate": 5.2222222222222226e-06, "loss": 0.1486, "step": 2650 }, { "epoch": 1.892465511142554, "grad_norm": 1.840253472328186, "learning_rate": 5.1666666666666675e-06, "loss": 0.1516, "step": 2675 }, { "epoch": 1.910152104704634, "grad_norm": 1.9791252613067627, "learning_rate": 5.1111111111111115e-06, "loss": 0.1431, "step": 2700 }, { "epoch": 1.9278386982667137, "grad_norm": 1.5719287395477295, "learning_rate": 5.0555555555555555e-06, "loss": 0.1539, "step": 2725 }, { "epoch": 1.9455252918287937, "grad_norm": 2.328838348388672, "learning_rate": 5e-06, "loss": 0.1502, "step": 2750 }, { "epoch": 1.9632118853908738, "grad_norm": 1.7131285667419434, "learning_rate": 4.944444444444445e-06, "loss": 0.1469, "step": 2775 }, { "epoch": 1.9808984789529536, "grad_norm": 1.751104712486267, "learning_rate": 4.888888888888889e-06, "loss": 0.151, "step": 2800 }, { "epoch": 1.9985850725150336, "grad_norm": 1.8015530109405518, "learning_rate": 4.833333333333333e-06, "loss": 0.1446, "step": 2825 }, { "epoch": 2.0162716660771136, "grad_norm": 1.4693965911865234, "learning_rate": 4.777777777777778e-06, "loss": 0.0913, "step": 2850 }, { "epoch": 2.0339582596391934, "grad_norm": 1.5922882556915283, "learning_rate": 4.722222222222222e-06, "loss": 0.0824, "step": 2875 }, { "epoch": 2.0516448532012737, "grad_norm": 1.0818850994110107, "learning_rate": 4.666666666666667e-06, "loss": 0.08, "step": 2900 }, { "epoch": 2.0693314467633535, "grad_norm": 1.5331037044525146, "learning_rate": 4.611111111111112e-06, "loss": 0.0761, "step": 2925 }, { "epoch": 2.0870180403254333, "grad_norm": 1.4420839548110962, "learning_rate": 4.555555555555556e-06, "loss": 0.072, "step": 2950 }, { "epoch": 2.104704633887513, "grad_norm": 1.4210786819458008, "learning_rate": 4.5e-06, "loss": 0.073, "step": 2975 }, { "epoch": 2.1223912274495933, "grad_norm": 1.2918660640716553, "learning_rate": 4.444444444444444e-06, "loss": 0.0714, "step": 3000 }, { "epoch": 2.1223912274495933, "eval_loss": 0.27894869446754456, "eval_runtime": 4698.3805, "eval_samples_per_second": 2.291, "eval_steps_per_second": 0.143, "eval_wer": 0.18066963936478134, "step": 3000 }, { "epoch": 2.140077821011673, "grad_norm": 1.2981915473937988, "learning_rate": 4.388888888888889e-06, "loss": 0.0705, "step": 3025 }, { "epoch": 2.157764414573753, "grad_norm": 1.5843642950057983, "learning_rate": 4.333333333333334e-06, "loss": 0.0801, "step": 3050 }, { "epoch": 2.175451008135833, "grad_norm": 1.4696415662765503, "learning_rate": 4.277777777777778e-06, "loss": 0.0828, "step": 3075 }, { "epoch": 2.193137601697913, "grad_norm": 1.4634337425231934, "learning_rate": 4.222222222222223e-06, "loss": 0.0744, "step": 3100 }, { "epoch": 2.2108241952599927, "grad_norm": 1.241773247718811, "learning_rate": 4.166666666666667e-06, "loss": 0.0735, "step": 3125 }, { "epoch": 2.228510788822073, "grad_norm": 1.7622851133346558, "learning_rate": 4.111111111111111e-06, "loss": 0.0755, "step": 3150 }, { "epoch": 2.246197382384153, "grad_norm": 1.7847949266433716, "learning_rate": 4.055555555555556e-06, "loss": 0.0697, "step": 3175 }, { "epoch": 2.2638839759462326, "grad_norm": 1.7188225984573364, "learning_rate": 4.000000000000001e-06, "loss": 0.0783, "step": 3200 }, { "epoch": 2.281570569508313, "grad_norm": 1.4745049476623535, "learning_rate": 3.944444444444445e-06, "loss": 0.0782, "step": 3225 }, { "epoch": 2.2992571630703926, "grad_norm": 1.488161563873291, "learning_rate": 3.88888888888889e-06, "loss": 0.081, "step": 3250 }, { "epoch": 2.3169437566324724, "grad_norm": 1.4803715944290161, "learning_rate": 3.833333333333334e-06, "loss": 0.0734, "step": 3275 }, { "epoch": 2.3346303501945527, "grad_norm": 1.7518547773361206, "learning_rate": 3.777777777777778e-06, "loss": 0.0791, "step": 3300 }, { "epoch": 2.3523169437566325, "grad_norm": 1.9586306810379028, "learning_rate": 3.7222222222222225e-06, "loss": 0.0814, "step": 3325 }, { "epoch": 2.3700035373187123, "grad_norm": 1.8565925359725952, "learning_rate": 3.6666666666666666e-06, "loss": 0.0728, "step": 3350 }, { "epoch": 2.3876901308807925, "grad_norm": 1.5781797170639038, "learning_rate": 3.6111111111111115e-06, "loss": 0.0826, "step": 3375 }, { "epoch": 2.4053767244428723, "grad_norm": 1.5568901300430298, "learning_rate": 3.555555555555556e-06, "loss": 0.0802, "step": 3400 }, { "epoch": 2.423063318004952, "grad_norm": 1.2662155628204346, "learning_rate": 3.5e-06, "loss": 0.074, "step": 3425 }, { "epoch": 2.4407499115670324, "grad_norm": 1.559577226638794, "learning_rate": 3.444444444444445e-06, "loss": 0.0758, "step": 3450 }, { "epoch": 2.458436505129112, "grad_norm": 1.5125735998153687, "learning_rate": 3.3888888888888893e-06, "loss": 0.0756, "step": 3475 }, { "epoch": 2.476123098691192, "grad_norm": 1.4088867902755737, "learning_rate": 3.3333333333333333e-06, "loss": 0.0713, "step": 3500 }, { "epoch": 2.493809692253272, "grad_norm": 1.6018534898757935, "learning_rate": 3.277777777777778e-06, "loss": 0.0656, "step": 3525 }, { "epoch": 2.511496285815352, "grad_norm": 1.5624502897262573, "learning_rate": 3.2222222222222227e-06, "loss": 0.0802, "step": 3550 }, { "epoch": 2.529182879377432, "grad_norm": 1.4694902896881104, "learning_rate": 3.1666666666666667e-06, "loss": 0.0713, "step": 3575 }, { "epoch": 2.5468694729395116, "grad_norm": 1.6186368465423584, "learning_rate": 3.1111111111111116e-06, "loss": 0.0864, "step": 3600 }, { "epoch": 2.564556066501592, "grad_norm": 1.6998704671859741, "learning_rate": 3.055555555555556e-06, "loss": 0.0721, "step": 3625 }, { "epoch": 2.5822426600636716, "grad_norm": 1.7324111461639404, "learning_rate": 3e-06, "loss": 0.0747, "step": 3650 }, { "epoch": 2.599929253625752, "grad_norm": 1.2067221403121948, "learning_rate": 2.944444444444445e-06, "loss": 0.0738, "step": 3675 }, { "epoch": 2.6176158471878317, "grad_norm": 1.8782434463500977, "learning_rate": 2.888888888888889e-06, "loss": 0.0699, "step": 3700 }, { "epoch": 2.6353024407499115, "grad_norm": 1.240626335144043, "learning_rate": 2.8333333333333335e-06, "loss": 0.0747, "step": 3725 }, { "epoch": 2.6529890343119913, "grad_norm": 1.2365247011184692, "learning_rate": 2.7777777777777783e-06, "loss": 0.0734, "step": 3750 }, { "epoch": 2.6706756278740715, "grad_norm": 1.5169546604156494, "learning_rate": 2.7222222222222224e-06, "loss": 0.0734, "step": 3775 }, { "epoch": 2.6883622214361513, "grad_norm": 1.3478069305419922, "learning_rate": 2.666666666666667e-06, "loss": 0.0757, "step": 3800 }, { "epoch": 2.7060488149982316, "grad_norm": 1.938991665840149, "learning_rate": 2.6111111111111113e-06, "loss": 0.0774, "step": 3825 }, { "epoch": 2.7237354085603114, "grad_norm": 1.7492825984954834, "learning_rate": 2.5555555555555557e-06, "loss": 0.0679, "step": 3850 }, { "epoch": 2.741422002122391, "grad_norm": 1.7618768215179443, "learning_rate": 2.5e-06, "loss": 0.0702, "step": 3875 }, { "epoch": 2.759108595684471, "grad_norm": 1.836369276046753, "learning_rate": 2.4444444444444447e-06, "loss": 0.0707, "step": 3900 }, { "epoch": 2.776795189246551, "grad_norm": 1.0155326128005981, "learning_rate": 2.388888888888889e-06, "loss": 0.0692, "step": 3925 }, { "epoch": 2.794481782808631, "grad_norm": 1.5357517004013062, "learning_rate": 2.3333333333333336e-06, "loss": 0.064, "step": 3950 }, { "epoch": 2.8121683763707113, "grad_norm": 1.413317322731018, "learning_rate": 2.277777777777778e-06, "loss": 0.0761, "step": 3975 }, { "epoch": 2.829854969932791, "grad_norm": 1.81116783618927, "learning_rate": 2.222222222222222e-06, "loss": 0.0772, "step": 4000 }, { "epoch": 2.829854969932791, "eval_loss": 0.27590474486351013, "eval_runtime": 4660.1057, "eval_samples_per_second": 2.31, "eval_steps_per_second": 0.144, "eval_wer": 0.18098255495580068, "step": 4000 }, { "epoch": 2.847541563494871, "grad_norm": 1.819291591644287, "learning_rate": 2.166666666666667e-06, "loss": 0.0739, "step": 4025 }, { "epoch": 2.8652281570569507, "grad_norm": 1.4449914693832397, "learning_rate": 2.1111111111111114e-06, "loss": 0.0672, "step": 4050 }, { "epoch": 2.882914750619031, "grad_norm": 1.617101788520813, "learning_rate": 2.0555555555555555e-06, "loss": 0.072, "step": 4075 }, { "epoch": 2.9006013441811107, "grad_norm": 1.7751729488372803, "learning_rate": 2.0000000000000003e-06, "loss": 0.075, "step": 4100 }, { "epoch": 2.9182879377431905, "grad_norm": 1.939385175704956, "learning_rate": 1.944444444444445e-06, "loss": 0.0746, "step": 4125 }, { "epoch": 2.9359745313052708, "grad_norm": 1.2767341136932373, "learning_rate": 1.888888888888889e-06, "loss": 0.065, "step": 4150 }, { "epoch": 2.9536611248673506, "grad_norm": 1.384352445602417, "learning_rate": 1.8333333333333333e-06, "loss": 0.0729, "step": 4175 }, { "epoch": 2.9713477184294304, "grad_norm": 1.707626223564148, "learning_rate": 1.777777777777778e-06, "loss": 0.0699, "step": 4200 }, { "epoch": 2.9890343119915106, "grad_norm": 1.5926012992858887, "learning_rate": 1.7222222222222224e-06, "loss": 0.0653, "step": 4225 }, { "epoch": 3.0067209055535904, "grad_norm": 0.7629415988922119, "learning_rate": 1.6666666666666667e-06, "loss": 0.0602, "step": 4250 }, { "epoch": 3.02440749911567, "grad_norm": 1.042358636856079, "learning_rate": 1.6111111111111113e-06, "loss": 0.0343, "step": 4275 }, { "epoch": 3.0420940926777504, "grad_norm": 1.4969948530197144, "learning_rate": 1.5555555555555558e-06, "loss": 0.0375, "step": 4300 }, { "epoch": 3.0597806862398302, "grad_norm": 2.0350911617279053, "learning_rate": 1.5e-06, "loss": 0.0343, "step": 4325 }, { "epoch": 3.07746727980191, "grad_norm": 1.1933727264404297, "learning_rate": 1.4444444444444445e-06, "loss": 0.0419, "step": 4350 }, { "epoch": 3.0951538733639903, "grad_norm": 1.2081432342529297, "learning_rate": 1.3888888888888892e-06, "loss": 0.0349, "step": 4375 }, { "epoch": 3.11284046692607, "grad_norm": 1.3141449689865112, "learning_rate": 1.3333333333333334e-06, "loss": 0.0357, "step": 4400 }, { "epoch": 3.13052706048815, "grad_norm": 1.156983494758606, "learning_rate": 1.2777777777777779e-06, "loss": 0.0313, "step": 4425 }, { "epoch": 3.14821365405023, "grad_norm": 0.8983932137489319, "learning_rate": 1.2222222222222223e-06, "loss": 0.0324, "step": 4450 }, { "epoch": 3.16590024761231, "grad_norm": 0.7461453676223755, "learning_rate": 1.1666666666666668e-06, "loss": 0.0414, "step": 4475 }, { "epoch": 3.1835868411743897, "grad_norm": 1.092787265777588, "learning_rate": 1.111111111111111e-06, "loss": 0.0363, "step": 4500 }, { "epoch": 3.2012734347364695, "grad_norm": 1.2066584825515747, "learning_rate": 1.0555555555555557e-06, "loss": 0.0361, "step": 4525 }, { "epoch": 3.2189600282985498, "grad_norm": 0.7660400867462158, "learning_rate": 1.0000000000000002e-06, "loss": 0.0338, "step": 4550 }, { "epoch": 3.2366466218606296, "grad_norm": 0.8779363632202148, "learning_rate": 9.444444444444445e-07, "loss": 0.0416, "step": 4575 }, { "epoch": 3.25433321542271, "grad_norm": 1.0976966619491577, "learning_rate": 8.88888888888889e-07, "loss": 0.0381, "step": 4600 }, { "epoch": 3.2720198089847896, "grad_norm": 0.977162778377533, "learning_rate": 8.333333333333333e-07, "loss": 0.0318, "step": 4625 }, { "epoch": 3.2897064025468694, "grad_norm": 1.5296571254730225, "learning_rate": 7.777777777777779e-07, "loss": 0.037, "step": 4650 }, { "epoch": 3.307392996108949, "grad_norm": 1.0584012269973755, "learning_rate": 7.222222222222222e-07, "loss": 0.0369, "step": 4675 }, { "epoch": 3.3250795896710295, "grad_norm": 0.8479712009429932, "learning_rate": 6.666666666666667e-07, "loss": 0.0343, "step": 4700 }, { "epoch": 3.3427661832331093, "grad_norm": 1.176857352256775, "learning_rate": 6.111111111111112e-07, "loss": 0.0349, "step": 4725 }, { "epoch": 3.360452776795189, "grad_norm": 1.1220146417617798, "learning_rate": 5.555555555555555e-07, "loss": 0.0307, "step": 4750 }, { "epoch": 3.3781393703572693, "grad_norm": 0.9692522883415222, "learning_rate": 5.000000000000001e-07, "loss": 0.0345, "step": 4775 }, { "epoch": 3.395825963919349, "grad_norm": 0.8425891995429993, "learning_rate": 4.444444444444445e-07, "loss": 0.0373, "step": 4800 }, { "epoch": 3.413512557481429, "grad_norm": 1.2310644388198853, "learning_rate": 3.8888888888888895e-07, "loss": 0.032, "step": 4825 }, { "epoch": 3.431199151043509, "grad_norm": 1.2488874197006226, "learning_rate": 3.3333333333333335e-07, "loss": 0.033, "step": 4850 }, { "epoch": 3.448885744605589, "grad_norm": 0.8309725522994995, "learning_rate": 2.7777777777777776e-07, "loss": 0.035, "step": 4875 }, { "epoch": 3.4665723381676687, "grad_norm": 0.8922958970069885, "learning_rate": 2.2222222222222224e-07, "loss": 0.0306, "step": 4900 }, { "epoch": 3.484258931729749, "grad_norm": 0.7373726963996887, "learning_rate": 1.6666666666666668e-07, "loss": 0.0324, "step": 4925 }, { "epoch": 3.501945525291829, "grad_norm": 1.3404922485351562, "learning_rate": 1.1111111111111112e-07, "loss": 0.0362, "step": 4950 }, { "epoch": 3.5196321188539086, "grad_norm": 1.187138319015503, "learning_rate": 5.555555555555556e-08, "loss": 0.0373, "step": 4975 }, { "epoch": 3.537318712415989, "grad_norm": 0.8413584232330322, "learning_rate": 0.0, "loss": 0.0337, "step": 5000 }, { "epoch": 3.537318712415989, "eval_loss": 0.3035752773284912, "eval_runtime": 4623.5314, "eval_samples_per_second": 2.328, "eval_steps_per_second": 0.146, "eval_wer": 0.18264491903309082, "step": 5000 }, { "epoch": 3.537318712415989, "step": 5000, "total_flos": 5.435079965953229e+20, "train_loss": 0.17054095619916915, "train_runtime": 54364.2112, "train_samples_per_second": 2.943, "train_steps_per_second": 0.092 } ], "logging_steps": 25, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.435079965953229e+20, "train_batch_size": 16, "trial_name": null, "trial_params": null }