{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 313, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0032, "grad_norm": 15.524961471557617, "learning_rate": 0.0, "loss": 1.9735, "step": 1 }, { "epoch": 0.0064, "grad_norm": 13.42629623413086, "learning_rate": 1.5625e-06, "loss": 1.7784, "step": 2 }, { "epoch": 0.0096, "grad_norm": 13.618926048278809, "learning_rate": 3.125e-06, "loss": 1.8501, "step": 3 }, { "epoch": 0.0128, "grad_norm": 13.631352424621582, "learning_rate": 4.6875000000000004e-06, "loss": 1.8382, "step": 4 }, { "epoch": 0.016, "grad_norm": 12.379663467407227, "learning_rate": 6.25e-06, "loss": 1.7936, "step": 5 }, { "epoch": 0.0192, "grad_norm": 9.427712440490723, "learning_rate": 7.8125e-06, "loss": 1.6236, "step": 6 }, { "epoch": 0.0224, "grad_norm": 8.059680938720703, "learning_rate": 9.375000000000001e-06, "loss": 1.4075, "step": 7 }, { "epoch": 0.0256, "grad_norm": 8.344525337219238, "learning_rate": 1.09375e-05, "loss": 1.5812, "step": 8 }, { "epoch": 0.0288, "grad_norm": 8.296407699584961, "learning_rate": 1.25e-05, "loss": 1.4497, "step": 9 }, { "epoch": 0.032, "grad_norm": 8.533961296081543, "learning_rate": 1.4062500000000001e-05, "loss": 1.3985, "step": 10 }, { "epoch": 0.0352, "grad_norm": 6.942614555358887, "learning_rate": 1.5625e-05, "loss": 1.3606, "step": 11 }, { "epoch": 0.0384, "grad_norm": 7.6024088859558105, "learning_rate": 1.71875e-05, "loss": 1.3757, "step": 12 }, { "epoch": 0.0416, "grad_norm": 7.377843856811523, "learning_rate": 1.8750000000000002e-05, "loss": 1.3528, "step": 13 }, { "epoch": 0.0448, "grad_norm": 7.739803791046143, "learning_rate": 2.0312500000000002e-05, "loss": 1.246, "step": 14 }, { "epoch": 0.048, "grad_norm": 6.864417552947998, "learning_rate": 2.1875e-05, "loss": 1.2315, "step": 15 }, { "epoch": 0.0512, "grad_norm": 7.203100681304932, "learning_rate": 2.34375e-05, "loss": 1.1748, "step": 16 }, { "epoch": 0.0544, "grad_norm": 8.64692497253418, "learning_rate": 2.5e-05, "loss": 1.1669, "step": 17 }, { "epoch": 0.0576, "grad_norm": 6.809157848358154, "learning_rate": 2.6562500000000002e-05, "loss": 1.0799, "step": 18 }, { "epoch": 0.0608, "grad_norm": 8.925545692443848, "learning_rate": 2.8125000000000003e-05, "loss": 1.0562, "step": 19 }, { "epoch": 0.064, "grad_norm": 5.894463539123535, "learning_rate": 2.96875e-05, "loss": 0.9397, "step": 20 }, { "epoch": 0.0672, "grad_norm": 4.393126010894775, "learning_rate": 3.125e-05, "loss": 0.9909, "step": 21 }, { "epoch": 0.0704, "grad_norm": 5.031020164489746, "learning_rate": 3.2812500000000005e-05, "loss": 0.9835, "step": 22 }, { "epoch": 0.0736, "grad_norm": 5.137537956237793, "learning_rate": 3.4375e-05, "loss": 1.011, "step": 23 }, { "epoch": 0.0768, "grad_norm": 4.404738903045654, "learning_rate": 3.59375e-05, "loss": 0.985, "step": 24 }, { "epoch": 0.08, "grad_norm": 4.291604518890381, "learning_rate": 3.7500000000000003e-05, "loss": 0.933, "step": 25 }, { "epoch": 0.0832, "grad_norm": 3.8908796310424805, "learning_rate": 3.90625e-05, "loss": 0.8894, "step": 26 }, { "epoch": 0.0864, "grad_norm": 5.403290271759033, "learning_rate": 4.0625000000000005e-05, "loss": 0.9313, "step": 27 }, { "epoch": 0.0896, "grad_norm": 4.857142448425293, "learning_rate": 4.21875e-05, "loss": 0.8802, "step": 28 }, { "epoch": 0.0928, "grad_norm": 5.22381067276001, "learning_rate": 4.375e-05, "loss": 0.925, "step": 29 }, { "epoch": 0.096, "grad_norm": 5.229774475097656, "learning_rate": 4.5312500000000004e-05, "loss": 0.9258, "step": 30 }, { "epoch": 0.0992, "grad_norm": 4.439915180206299, "learning_rate": 4.6875e-05, "loss": 0.8377, "step": 31 }, { "epoch": 0.1024, "grad_norm": 4.758205413818359, "learning_rate": 4.8437500000000005e-05, "loss": 0.8652, "step": 32 }, { "epoch": 0.1056, "grad_norm": 4.30954122543335, "learning_rate": 5e-05, "loss": 0.884, "step": 33 }, { "epoch": 0.1088, "grad_norm": 4.055234909057617, "learning_rate": 4.999843759868819e-05, "loss": 0.7903, "step": 34 }, { "epoch": 0.112, "grad_norm": 5.193882942199707, "learning_rate": 4.9993750590040575e-05, "loss": 0.8985, "step": 35 }, { "epoch": 0.1152, "grad_norm": 4.579808235168457, "learning_rate": 4.998593955989626e-05, "loss": 0.8734, "step": 36 }, { "epoch": 0.1184, "grad_norm": 3.9402709007263184, "learning_rate": 4.9975005484572305e-05, "loss": 0.8567, "step": 37 }, { "epoch": 0.1216, "grad_norm": 4.5101704597473145, "learning_rate": 4.996094973074183e-05, "loss": 0.9164, "step": 38 }, { "epoch": 0.1248, "grad_norm": 4.826318264007568, "learning_rate": 4.994377405526308e-05, "loss": 0.8821, "step": 39 }, { "epoch": 0.128, "grad_norm": 3.9617154598236084, "learning_rate": 4.992348060495989e-05, "loss": 0.883, "step": 40 }, { "epoch": 0.1312, "grad_norm": 3.356626272201538, "learning_rate": 4.990007191635334e-05, "loss": 0.7852, "step": 41 }, { "epoch": 0.1344, "grad_norm": 3.9243972301483154, "learning_rate": 4.987355091534468e-05, "loss": 0.8668, "step": 42 }, { "epoch": 0.1376, "grad_norm": 3.6170945167541504, "learning_rate": 4.9843920916849645e-05, "loss": 0.8517, "step": 43 }, { "epoch": 0.1408, "grad_norm": 4.146400451660156, "learning_rate": 4.981118562438414e-05, "loss": 0.9207, "step": 44 }, { "epoch": 0.144, "grad_norm": 4.426783561706543, "learning_rate": 4.9775349129601243e-05, "loss": 0.8918, "step": 45 }, { "epoch": 0.1472, "grad_norm": 3.632885217666626, "learning_rate": 4.973641591177991e-05, "loss": 0.8404, "step": 46 }, { "epoch": 0.1504, "grad_norm": 4.035740375518799, "learning_rate": 4.969439083726496e-05, "loss": 0.8945, "step": 47 }, { "epoch": 0.1536, "grad_norm": 3.7760114669799805, "learning_rate": 4.964927915885893e-05, "loss": 0.7685, "step": 48 }, { "epoch": 0.1568, "grad_norm": 3.9631412029266357, "learning_rate": 4.960108651516545e-05, "loss": 0.8538, "step": 49 }, { "epoch": 0.16, "grad_norm": 3.716055393218994, "learning_rate": 4.954981892988451e-05, "loss": 0.8328, "step": 50 }, { "epoch": 0.1632, "grad_norm": 4.943089485168457, "learning_rate": 4.949548281105951e-05, "loss": 0.8006, "step": 51 }, { "epoch": 0.1664, "grad_norm": 3.410830497741699, "learning_rate": 4.943808495027631e-05, "loss": 0.8173, "step": 52 }, { "epoch": 0.1696, "grad_norm": 3.5364482402801514, "learning_rate": 4.937763252181434e-05, "loss": 0.8017, "step": 53 }, { "epoch": 0.1728, "grad_norm": 3.6835520267486572, "learning_rate": 4.93141330817499e-05, "loss": 0.8256, "step": 54 }, { "epoch": 0.176, "grad_norm": 3.966698408126831, "learning_rate": 4.924759456701167e-05, "loss": 0.8032, "step": 55 }, { "epoch": 0.1792, "grad_norm": 4.649932384490967, "learning_rate": 4.917802529438864e-05, "loss": 0.9344, "step": 56 }, { "epoch": 0.1824, "grad_norm": 3.5674755573272705, "learning_rate": 4.910543395949067e-05, "loss": 0.8222, "step": 57 }, { "epoch": 0.1856, "grad_norm": 4.073423862457275, "learning_rate": 4.9029829635661475e-05, "loss": 0.8949, "step": 58 }, { "epoch": 0.1888, "grad_norm": 4.209728717803955, "learning_rate": 4.895122177284465e-05, "loss": 0.8474, "step": 59 }, { "epoch": 0.192, "grad_norm": 4.795137882232666, "learning_rate": 4.8869620196402436e-05, "loss": 0.861, "step": 60 }, { "epoch": 0.1952, "grad_norm": 3.300403594970703, "learning_rate": 4.878503510588765e-05, "loss": 0.828, "step": 61 }, { "epoch": 0.1984, "grad_norm": 3.2910377979278564, "learning_rate": 4.8697477073768766e-05, "loss": 0.8032, "step": 62 }, { "epoch": 0.2016, "grad_norm": 3.4420082569122314, "learning_rate": 4.8606957044108556e-05, "loss": 0.7789, "step": 63 }, { "epoch": 0.2048, "grad_norm": 3.449355363845825, "learning_rate": 4.851348633119606e-05, "loss": 0.6904, "step": 64 }, { "epoch": 0.208, "grad_norm": 3.5272228717803955, "learning_rate": 4.8417076618132426e-05, "loss": 0.7907, "step": 65 }, { "epoch": 0.2112, "grad_norm": 3.2754251956939697, "learning_rate": 4.8317739955370636e-05, "loss": 0.7932, "step": 66 }, { "epoch": 0.2144, "grad_norm": 3.522357702255249, "learning_rate": 4.821548875920927e-05, "loss": 0.7782, "step": 67 }, { "epoch": 0.2176, "grad_norm": 3.652512311935425, "learning_rate": 4.811033581024056e-05, "loss": 0.7733, "step": 68 }, { "epoch": 0.2208, "grad_norm": 3.8177380561828613, "learning_rate": 4.800229425175294e-05, "loss": 0.7116, "step": 69 }, { "epoch": 0.224, "grad_norm": 4.430614948272705, "learning_rate": 4.7891377588088223e-05, "loss": 0.8089, "step": 70 }, { "epoch": 0.2272, "grad_norm": 4.25429630279541, "learning_rate": 4.777759968295369e-05, "loss": 0.7455, "step": 71 }, { "epoch": 0.2304, "grad_norm": 3.8671886920928955, "learning_rate": 4.766097475768919e-05, "loss": 0.8566, "step": 72 }, { "epoch": 0.2336, "grad_norm": 4.336380481719971, "learning_rate": 4.754151738948962e-05, "loss": 0.8277, "step": 73 }, { "epoch": 0.2368, "grad_norm": 5.870267391204834, "learning_rate": 4.741924250958289e-05, "loss": 0.7712, "step": 74 }, { "epoch": 0.24, "grad_norm": 5.58464241027832, "learning_rate": 4.729416540136361e-05, "loss": 0.837, "step": 75 }, { "epoch": 0.2432, "grad_norm": 6.278571128845215, "learning_rate": 4.7166301698482815e-05, "loss": 0.7804, "step": 76 }, { "epoch": 0.2464, "grad_norm": 3.8727006912231445, "learning_rate": 4.703566738289389e-05, "loss": 0.7113, "step": 77 }, { "epoch": 0.2496, "grad_norm": 2.9273529052734375, "learning_rate": 4.69022787828549e-05, "loss": 0.7402, "step": 78 }, { "epoch": 0.2528, "grad_norm": 4.868607997894287, "learning_rate": 4.676615257088776e-05, "loss": 0.7648, "step": 79 }, { "epoch": 0.256, "grad_norm": 4.873749732971191, "learning_rate": 4.662730576169423e-05, "loss": 0.8317, "step": 80 }, { "epoch": 0.2592, "grad_norm": 3.7857329845428467, "learning_rate": 4.6485755710029256e-05, "loss": 0.7691, "step": 81 }, { "epoch": 0.2624, "grad_norm": 4.516591548919678, "learning_rate": 4.6341520108531746e-05, "loss": 0.8074, "step": 82 }, { "epoch": 0.2656, "grad_norm": 2.922186851501465, "learning_rate": 4.619461698551315e-05, "loss": 0.6979, "step": 83 }, { "epoch": 0.2688, "grad_norm": 3.1487746238708496, "learning_rate": 4.604506470270403e-05, "loss": 0.8106, "step": 84 }, { "epoch": 0.272, "grad_norm": 3.4952852725982666, "learning_rate": 4.589288195295901e-05, "loss": 0.8193, "step": 85 }, { "epoch": 0.2752, "grad_norm": 3.446680784225464, "learning_rate": 4.573808775792033e-05, "loss": 0.6519, "step": 86 }, { "epoch": 0.2784, "grad_norm": 4.01785135269165, "learning_rate": 4.5580701465640254e-05, "loss": 0.7972, "step": 87 }, { "epoch": 0.2816, "grad_norm": 3.0481715202331543, "learning_rate": 4.5420742748162734e-05, "loss": 0.8058, "step": 88 }, { "epoch": 0.2848, "grad_norm": 3.555839776992798, "learning_rate": 4.525823159906459e-05, "loss": 0.8275, "step": 89 }, { "epoch": 0.288, "grad_norm": 3.641484260559082, "learning_rate": 4.509318833095642e-05, "loss": 0.7904, "step": 90 }, { "epoch": 0.2912, "grad_norm": 3.2470881938934326, "learning_rate": 4.492563357294369e-05, "loss": 0.7142, "step": 91 }, { "epoch": 0.2944, "grad_norm": 2.9820683002471924, "learning_rate": 4.475558826804833e-05, "loss": 0.7862, "step": 92 }, { "epoch": 0.2976, "grad_norm": 3.468200206756592, "learning_rate": 4.458307367059092e-05, "loss": 0.7648, "step": 93 }, { "epoch": 0.3008, "grad_norm": 3.363412380218506, "learning_rate": 4.440811134353412e-05, "loss": 0.7217, "step": 94 }, { "epoch": 0.304, "grad_norm": 3.7876627445220947, "learning_rate": 4.42307231557875e-05, "loss": 0.7541, "step": 95 }, { "epoch": 0.3072, "grad_norm": 3.116093397140503, "learning_rate": 4.4050931279474015e-05, "loss": 0.7603, "step": 96 }, { "epoch": 0.3104, "grad_norm": 3.113248109817505, "learning_rate": 4.386875818715874e-05, "loss": 0.6, "step": 97 }, { "epoch": 0.3136, "grad_norm": 3.3155691623687744, "learning_rate": 4.368422664903997e-05, "loss": 0.7264, "step": 98 }, { "epoch": 0.3168, "grad_norm": 4.281957626342773, "learning_rate": 4.349735973010305e-05, "loss": 0.8178, "step": 99 }, { "epoch": 0.32, "grad_norm": 4.118067264556885, "learning_rate": 4.330818078723755e-05, "loss": 0.7394, "step": 100 }, { "epoch": 0.3232, "grad_norm": 3.7865848541259766, "learning_rate": 4.311671346631774e-05, "loss": 0.8166, "step": 101 }, { "epoch": 0.3264, "grad_norm": 3.1080141067504883, "learning_rate": 4.292298169924709e-05, "loss": 0.6196, "step": 102 }, { "epoch": 0.3296, "grad_norm": 2.8744795322418213, "learning_rate": 4.272700970096696e-05, "loss": 0.7371, "step": 103 }, { "epoch": 0.3328, "grad_norm": 3.6204240322113037, "learning_rate": 4.252882196642992e-05, "loss": 0.7253, "step": 104 }, { "epoch": 0.336, "grad_norm": 2.7233681678771973, "learning_rate": 4.23284432675381e-05, "loss": 0.6528, "step": 105 }, { "epoch": 0.3392, "grad_norm": 2.825315237045288, "learning_rate": 4.212589865004684e-05, "loss": 0.6597, "step": 106 }, { "epoch": 0.3424, "grad_norm": 3.6910030841827393, "learning_rate": 4.192121343043424e-05, "loss": 0.6703, "step": 107 }, { "epoch": 0.3456, "grad_norm": 3.4581615924835205, "learning_rate": 4.1714413192736754e-05, "loss": 0.759, "step": 108 }, { "epoch": 0.3488, "grad_norm": 4.329128265380859, "learning_rate": 4.150552378535137e-05, "loss": 0.7174, "step": 109 }, { "epoch": 0.352, "grad_norm": 3.2338504791259766, "learning_rate": 4.1294571317804854e-05, "loss": 0.7708, "step": 110 }, { "epoch": 0.3552, "grad_norm": 3.127150774002075, "learning_rate": 4.108158215749014e-05, "loss": 0.6826, "step": 111 }, { "epoch": 0.3584, "grad_norm": 3.5068857669830322, "learning_rate": 4.0866582926370725e-05, "loss": 0.6749, "step": 112 }, { "epoch": 0.3616, "grad_norm": 3.0821373462677, "learning_rate": 4.064960049765304e-05, "loss": 0.7902, "step": 113 }, { "epoch": 0.3648, "grad_norm": 3.2901744842529297, "learning_rate": 4.043066199242762e-05, "loss": 0.7017, "step": 114 }, { "epoch": 0.368, "grad_norm": 3.2184112071990967, "learning_rate": 4.020979477627907e-05, "loss": 0.7647, "step": 115 }, { "epoch": 0.3712, "grad_norm": 3.092893600463867, "learning_rate": 3.998702645586565e-05, "loss": 0.6897, "step": 116 }, { "epoch": 0.3744, "grad_norm": 3.3295419216156006, "learning_rate": 3.976238487546864e-05, "loss": 0.8187, "step": 117 }, { "epoch": 0.3776, "grad_norm": 2.6530091762542725, "learning_rate": 3.953589811351204e-05, "loss": 0.7974, "step": 118 }, { "epoch": 0.3808, "grad_norm": 3.290668487548828, "learning_rate": 3.930759447905298e-05, "loss": 0.7172, "step": 119 }, { "epoch": 0.384, "grad_norm": 2.70243763923645, "learning_rate": 3.907750250824327e-05, "loss": 0.6843, "step": 120 }, { "epoch": 0.3872, "grad_norm": 3.1110892295837402, "learning_rate": 3.884565096076269e-05, "loss": 0.7896, "step": 121 }, { "epoch": 0.3904, "grad_norm": 2.7878258228302, "learning_rate": 3.861206881622419e-05, "loss": 0.7505, "step": 122 }, { "epoch": 0.3936, "grad_norm": 3.118669271469116, "learning_rate": 3.837678527055168e-05, "loss": 0.6734, "step": 123 }, { "epoch": 0.3968, "grad_norm": 3.1683592796325684, "learning_rate": 3.813982973233083e-05, "loss": 0.7451, "step": 124 }, { "epoch": 0.4, "grad_norm": 4.442941665649414, "learning_rate": 3.7901231819133105e-05, "loss": 0.8121, "step": 125 }, { "epoch": 0.4032, "grad_norm": 2.806699275970459, "learning_rate": 3.766102135381393e-05, "loss": 0.6691, "step": 126 }, { "epoch": 0.4064, "grad_norm": 3.6106672286987305, "learning_rate": 3.741922836078499e-05, "loss": 0.6728, "step": 127 }, { "epoch": 0.4096, "grad_norm": 3.241565465927124, "learning_rate": 3.717588306226143e-05, "loss": 0.7116, "step": 128 }, { "epoch": 0.4128, "grad_norm": 2.966460704803467, "learning_rate": 3.693101587448436e-05, "loss": 0.6842, "step": 129 }, { "epoch": 0.416, "grad_norm": 2.9147496223449707, "learning_rate": 3.6684657403919005e-05, "loss": 0.725, "step": 130 }, { "epoch": 0.4192, "grad_norm": 3.226004123687744, "learning_rate": 3.6436838443429175e-05, "loss": 0.6485, "step": 131 }, { "epoch": 0.4224, "grad_norm": 4.062314510345459, "learning_rate": 3.618758996842839e-05, "loss": 0.7387, "step": 132 }, { "epoch": 0.4256, "grad_norm": 2.843627691268921, "learning_rate": 3.5936943133008183e-05, "loss": 0.6621, "step": 133 }, { "epoch": 0.4288, "grad_norm": 3.151275396347046, "learning_rate": 3.568492926604412e-05, "loss": 0.733, "step": 134 }, { "epoch": 0.432, "grad_norm": 2.40847110748291, "learning_rate": 3.5431579867279905e-05, "loss": 0.6433, "step": 135 }, { "epoch": 0.4352, "grad_norm": 3.5088706016540527, "learning_rate": 3.517692660339018e-05, "loss": 0.8031, "step": 136 }, { "epoch": 0.4384, "grad_norm": 2.9651691913604736, "learning_rate": 3.492100130402242e-05, "loss": 0.6583, "step": 137 }, { "epoch": 0.4416, "grad_norm": 2.8937885761260986, "learning_rate": 3.4663835957818515e-05, "loss": 0.6887, "step": 138 }, { "epoch": 0.4448, "grad_norm": 3.493434429168701, "learning_rate": 3.440546270841639e-05, "loss": 0.7076, "step": 139 }, { "epoch": 0.448, "grad_norm": 2.6669135093688965, "learning_rate": 3.414591385043237e-05, "loss": 0.7377, "step": 140 }, { "epoch": 0.4512, "grad_norm": 3.7485995292663574, "learning_rate": 3.3885221825424537e-05, "loss": 0.6898, "step": 141 }, { "epoch": 0.4544, "grad_norm": 2.8648860454559326, "learning_rate": 3.362341921783784e-05, "loss": 0.7099, "step": 142 }, { "epoch": 0.4576, "grad_norm": 3.085545063018799, "learning_rate": 3.336053875093128e-05, "loss": 0.6491, "step": 143 }, { "epoch": 0.4608, "grad_norm": 2.9615533351898193, "learning_rate": 3.309661328268776e-05, "loss": 0.6848, "step": 144 }, { "epoch": 0.464, "grad_norm": 2.989152431488037, "learning_rate": 3.283167580170712e-05, "loss": 0.753, "step": 145 }, { "epoch": 0.4672, "grad_norm": 2.6297624111175537, "learning_rate": 3.256575942308278e-05, "loss": 0.6575, "step": 146 }, { "epoch": 0.4704, "grad_norm": 3.077404022216797, "learning_rate": 3.229889738426264e-05, "loss": 0.7943, "step": 147 }, { "epoch": 0.4736, "grad_norm": 3.0161232948303223, "learning_rate": 3.203112304089466e-05, "loss": 0.665, "step": 148 }, { "epoch": 0.4768, "grad_norm": 3.0550014972686768, "learning_rate": 3.176246986265767e-05, "loss": 0.7091, "step": 149 }, { "epoch": 0.48, "grad_norm": 3.3446528911590576, "learning_rate": 3.149297142907792e-05, "loss": 0.7586, "step": 150 }, { "epoch": 0.4832, "grad_norm": 2.50750470161438, "learning_rate": 3.122266142533191e-05, "loss": 0.6006, "step": 151 }, { "epoch": 0.4864, "grad_norm": 2.2805991172790527, "learning_rate": 3.095157363803598e-05, "loss": 0.6918, "step": 152 }, { "epoch": 0.4896, "grad_norm": 2.982093572616577, "learning_rate": 3.06797419510233e-05, "loss": 0.6261, "step": 153 }, { "epoch": 0.4928, "grad_norm": 3.545555591583252, "learning_rate": 3.0407200341108617e-05, "loss": 0.5964, "step": 154 }, { "epoch": 0.496, "grad_norm": 3.1337227821350098, "learning_rate": 3.013398287384144e-05, "loss": 0.6798, "step": 155 }, { "epoch": 0.4992, "grad_norm": 3.5548670291900635, "learning_rate": 2.986012369924811e-05, "loss": 0.72, "step": 156 }, { "epoch": 0.5024, "grad_norm": 2.6750857830047607, "learning_rate": 2.9585657047563315e-05, "loss": 0.7358, "step": 157 }, { "epoch": 0.5056, "grad_norm": 3.0403218269348145, "learning_rate": 2.931061722495159e-05, "loss": 0.6558, "step": 158 }, { "epoch": 0.5088, "grad_norm": 2.5860700607299805, "learning_rate": 2.9035038609219306e-05, "loss": 0.6607, "step": 159 }, { "epoch": 0.512, "grad_norm": 2.983950138092041, "learning_rate": 2.875895564551772e-05, "loss": 0.7403, "step": 160 }, { "epoch": 0.5152, "grad_norm": 2.4601857662200928, "learning_rate": 2.8482402842037614e-05, "loss": 0.581, "step": 161 }, { "epoch": 0.5184, "grad_norm": 2.6015350818634033, "learning_rate": 2.8205414765696003e-05, "loss": 0.6462, "step": 162 }, { "epoch": 0.5216, "grad_norm": 3.2562873363494873, "learning_rate": 2.792802603781562e-05, "loss": 0.6778, "step": 163 }, { "epoch": 0.5248, "grad_norm": 3.0855162143707275, "learning_rate": 2.7650271329797427e-05, "loss": 0.6187, "step": 164 }, { "epoch": 0.528, "grad_norm": 2.281909942626953, "learning_rate": 2.737218535878705e-05, "loss": 0.6209, "step": 165 }, { "epoch": 0.5312, "grad_norm": 2.689497232437134, "learning_rate": 2.7093802883335357e-05, "loss": 0.6529, "step": 166 }, { "epoch": 0.5344, "grad_norm": 2.919034719467163, "learning_rate": 2.6815158699053932e-05, "loss": 0.6751, "step": 167 }, { "epoch": 0.5376, "grad_norm": 2.32546067237854, "learning_rate": 2.6536287634265918e-05, "loss": 0.5795, "step": 168 }, { "epoch": 0.5408, "grad_norm": 2.528740167617798, "learning_rate": 2.6257224545652688e-05, "loss": 0.6675, "step": 169 }, { "epoch": 0.544, "grad_norm": 2.651212692260742, "learning_rate": 2.5978004313897104e-05, "loss": 0.6048, "step": 170 }, { "epoch": 0.5472, "grad_norm": 2.5883514881134033, "learning_rate": 2.569866183932368e-05, "loss": 0.584, "step": 171 }, { "epoch": 0.5504, "grad_norm": 2.2905187606811523, "learning_rate": 2.5419232037536316e-05, "loss": 0.6269, "step": 172 }, { "epoch": 0.5536, "grad_norm": 3.2044761180877686, "learning_rate": 2.5139749835054123e-05, "loss": 0.6114, "step": 173 }, { "epoch": 0.5568, "grad_norm": 2.599435806274414, "learning_rate": 2.4860250164945876e-05, "loss": 0.6962, "step": 174 }, { "epoch": 0.56, "grad_norm": 2.6137402057647705, "learning_rate": 2.4580767962463687e-05, "loss": 0.614, "step": 175 }, { "epoch": 0.5632, "grad_norm": 2.9657676219940186, "learning_rate": 2.4301338160676324e-05, "loss": 0.6616, "step": 176 }, { "epoch": 0.5664, "grad_norm": 2.5792715549468994, "learning_rate": 2.40219956861029e-05, "loss": 0.6584, "step": 177 }, { "epoch": 0.5696, "grad_norm": 2.874328851699829, "learning_rate": 2.374277545434732e-05, "loss": 0.6192, "step": 178 }, { "epoch": 0.5728, "grad_norm": 2.860802412033081, "learning_rate": 2.346371236573409e-05, "loss": 0.7215, "step": 179 }, { "epoch": 0.576, "grad_norm": 2.5315871238708496, "learning_rate": 2.318484130094607e-05, "loss": 0.6511, "step": 180 }, { "epoch": 0.5792, "grad_norm": 2.4689619541168213, "learning_rate": 2.2906197116664653e-05, "loss": 0.5986, "step": 181 }, { "epoch": 0.5824, "grad_norm": 2.6381800174713135, "learning_rate": 2.262781464121296e-05, "loss": 0.6328, "step": 182 }, { "epoch": 0.5856, "grad_norm": 2.6276397705078125, "learning_rate": 2.2349728670202582e-05, "loss": 0.6409, "step": 183 }, { "epoch": 0.5888, "grad_norm": 2.783932685852051, "learning_rate": 2.2071973962184384e-05, "loss": 0.6556, "step": 184 }, { "epoch": 0.592, "grad_norm": 3.184654474258423, "learning_rate": 2.1794585234303993e-05, "loss": 0.6706, "step": 185 }, { "epoch": 0.5952, "grad_norm": 2.7625272274017334, "learning_rate": 2.1517597157962392e-05, "loss": 0.5842, "step": 186 }, { "epoch": 0.5984, "grad_norm": 2.9620823860168457, "learning_rate": 2.124104435448228e-05, "loss": 0.622, "step": 187 }, { "epoch": 0.6016, "grad_norm": 3.4759445190429688, "learning_rate": 2.0964961390780703e-05, "loss": 0.6734, "step": 188 }, { "epoch": 0.6048, "grad_norm": 2.639575481414795, "learning_rate": 2.0689382775048418e-05, "loss": 0.6005, "step": 189 }, { "epoch": 0.608, "grad_norm": 2.5835771560668945, "learning_rate": 2.0414342952436694e-05, "loss": 0.6009, "step": 190 }, { "epoch": 0.6112, "grad_norm": 2.8783106803894043, "learning_rate": 2.0139876300751904e-05, "loss": 0.6493, "step": 191 }, { "epoch": 0.6144, "grad_norm": 2.6388046741485596, "learning_rate": 1.9866017126158574e-05, "loss": 0.591, "step": 192 }, { "epoch": 0.6176, "grad_norm": 2.5275888442993164, "learning_rate": 1.9592799658891385e-05, "loss": 0.5525, "step": 193 }, { "epoch": 0.6208, "grad_norm": 2.582974672317505, "learning_rate": 1.9320258048976702e-05, "loss": 0.5525, "step": 194 }, { "epoch": 0.624, "grad_norm": 2.4387340545654297, "learning_rate": 1.904842636196402e-05, "loss": 0.6502, "step": 195 }, { "epoch": 0.6272, "grad_norm": 3.4080708026885986, "learning_rate": 1.8777338574668095e-05, "loss": 0.5925, "step": 196 }, { "epoch": 0.6304, "grad_norm": 2.7693426609039307, "learning_rate": 1.850702857092208e-05, "loss": 0.6189, "step": 197 }, { "epoch": 0.6336, "grad_norm": 2.9749882221221924, "learning_rate": 1.8237530137342335e-05, "loss": 0.6079, "step": 198 }, { "epoch": 0.6368, "grad_norm": 2.405189275741577, "learning_rate": 1.796887695910535e-05, "loss": 0.4882, "step": 199 }, { "epoch": 0.64, "grad_norm": 2.856356143951416, "learning_rate": 1.7701102615737368e-05, "loss": 0.662, "step": 200 }, { "epoch": 0.6432, "grad_norm": 3.5571553707122803, "learning_rate": 1.7434240576917226e-05, "loss": 0.6705, "step": 201 }, { "epoch": 0.6464, "grad_norm": 2.9038398265838623, "learning_rate": 1.7168324198292888e-05, "loss": 0.647, "step": 202 }, { "epoch": 0.6496, "grad_norm": 2.6858866214752197, "learning_rate": 1.6903386717312236e-05, "loss": 0.5854, "step": 203 }, { "epoch": 0.6528, "grad_norm": 2.881903648376465, "learning_rate": 1.6639461249068726e-05, "loss": 0.6802, "step": 204 }, { "epoch": 0.656, "grad_norm": 2.7893338203430176, "learning_rate": 1.637658078216217e-05, "loss": 0.6101, "step": 205 }, { "epoch": 0.6592, "grad_norm": 2.7626793384552, "learning_rate": 1.6114778174575473e-05, "loss": 0.5851, "step": 206 }, { "epoch": 0.6624, "grad_norm": 2.4826924800872803, "learning_rate": 1.585408614956763e-05, "loss": 0.5924, "step": 207 }, { "epoch": 0.6656, "grad_norm": 2.7100701332092285, "learning_rate": 1.559453729158361e-05, "loss": 0.5931, "step": 208 }, { "epoch": 0.6688, "grad_norm": 2.625746488571167, "learning_rate": 1.5336164042181494e-05, "loss": 0.5665, "step": 209 }, { "epoch": 0.672, "grad_norm": 2.8717494010925293, "learning_rate": 1.5078998695977586e-05, "loss": 0.6568, "step": 210 }, { "epoch": 0.6752, "grad_norm": 2.610119581222534, "learning_rate": 1.482307339660983e-05, "loss": 0.5354, "step": 211 }, { "epoch": 0.6784, "grad_norm": 2.783958911895752, "learning_rate": 1.4568420132720106e-05, "loss": 0.5658, "step": 212 }, { "epoch": 0.6816, "grad_norm": 2.3880062103271484, "learning_rate": 1.4315070733955888e-05, "loss": 0.5361, "step": 213 }, { "epoch": 0.6848, "grad_norm": 2.3239946365356445, "learning_rate": 1.4063056866991826e-05, "loss": 0.5477, "step": 214 }, { "epoch": 0.688, "grad_norm": 3.000854730606079, "learning_rate": 1.381241003157162e-05, "loss": 0.5555, "step": 215 }, { "epoch": 0.6912, "grad_norm": 2.6521692276000977, "learning_rate": 1.3563161556570826e-05, "loss": 0.5702, "step": 216 }, { "epoch": 0.6944, "grad_norm": 2.789949417114258, "learning_rate": 1.3315342596080996e-05, "loss": 0.5894, "step": 217 }, { "epoch": 0.6976, "grad_norm": 2.630718231201172, "learning_rate": 1.3068984125515644e-05, "loss": 0.5601, "step": 218 }, { "epoch": 0.7008, "grad_norm": 2.7398900985717773, "learning_rate": 1.2824116937738579e-05, "loss": 0.6077, "step": 219 }, { "epoch": 0.704, "grad_norm": 2.9107697010040283, "learning_rate": 1.2580771639215027e-05, "loss": 0.5389, "step": 220 }, { "epoch": 0.7072, "grad_norm": 5.506183624267578, "learning_rate": 1.2338978646186084e-05, "loss": 0.565, "step": 221 }, { "epoch": 0.7104, "grad_norm": 2.896214008331299, "learning_rate": 1.2098768180866895e-05, "loss": 0.5993, "step": 222 }, { "epoch": 0.7136, "grad_norm": 2.5421102046966553, "learning_rate": 1.1860170267669174e-05, "loss": 0.6117, "step": 223 }, { "epoch": 0.7168, "grad_norm": 3.075676679611206, "learning_rate": 1.1623214729448317e-05, "loss": 0.5713, "step": 224 }, { "epoch": 0.72, "grad_norm": 2.6561267375946045, "learning_rate": 1.1387931183775822e-05, "loss": 0.553, "step": 225 }, { "epoch": 0.7232, "grad_norm": 2.5555975437164307, "learning_rate": 1.1154349039237322e-05, "loss": 0.5496, "step": 226 }, { "epoch": 0.7264, "grad_norm": 2.842125177383423, "learning_rate": 1.0922497491756734e-05, "loss": 0.6698, "step": 227 }, { "epoch": 0.7296, "grad_norm": 3.4003820419311523, "learning_rate": 1.0692405520947028e-05, "loss": 0.6218, "step": 228 }, { "epoch": 0.7328, "grad_norm": 2.400970697402954, "learning_rate": 1.0464101886487958e-05, "loss": 0.5709, "step": 229 }, { "epoch": 0.736, "grad_norm": 2.4426651000976562, "learning_rate": 1.0237615124531363e-05, "loss": 0.5921, "step": 230 }, { "epoch": 0.7392, "grad_norm": 3.67607045173645, "learning_rate": 1.0012973544134358e-05, "loss": 0.5904, "step": 231 }, { "epoch": 0.7424, "grad_norm": 2.6775803565979004, "learning_rate": 9.79020522372093e-06, "loss": 0.6506, "step": 232 }, { "epoch": 0.7456, "grad_norm": 2.234372854232788, "learning_rate": 9.569338007572382e-06, "loss": 0.5061, "step": 233 }, { "epoch": 0.7488, "grad_norm": 3.26226806640625, "learning_rate": 9.35039950234696e-06, "loss": 0.5994, "step": 234 }, { "epoch": 0.752, "grad_norm": 2.563432216644287, "learning_rate": 9.133417073629289e-06, "loss": 0.5684, "step": 235 }, { "epoch": 0.7552, "grad_norm": 2.5276975631713867, "learning_rate": 8.918417842509867e-06, "loss": 0.5555, "step": 236 }, { "epoch": 0.7584, "grad_norm": 2.4704692363739014, "learning_rate": 8.705428682195155e-06, "loss": 0.5411, "step": 237 }, { "epoch": 0.7616, "grad_norm": 2.7508208751678467, "learning_rate": 8.494476214648626e-06, "loss": 0.6805, "step": 238 }, { "epoch": 0.7648, "grad_norm": 2.4423398971557617, "learning_rate": 8.285586807263254e-06, "loss": 0.5083, "step": 239 }, { "epoch": 0.768, "grad_norm": 3.0647759437561035, "learning_rate": 8.078786569565763e-06, "loss": 0.6004, "step": 240 }, { "epoch": 0.7712, "grad_norm": 2.466031312942505, "learning_rate": 7.874101349953167e-06, "loss": 0.5198, "step": 241 }, { "epoch": 0.7744, "grad_norm": 2.9429163932800293, "learning_rate": 7.671556732461905e-06, "loss": 0.5538, "step": 242 }, { "epoch": 0.7776, "grad_norm": 2.379636526107788, "learning_rate": 7.471178033570081e-06, "loss": 0.4544, "step": 243 }, { "epoch": 0.7808, "grad_norm": 2.734103202819824, "learning_rate": 7.272990299033045e-06, "loss": 0.541, "step": 244 }, { "epoch": 0.784, "grad_norm": 2.27902889251709, "learning_rate": 7.077018300752916e-06, "loss": 0.5235, "step": 245 }, { "epoch": 0.7872, "grad_norm": 2.284369468688965, "learning_rate": 6.883286533682265e-06, "loss": 0.49, "step": 246 }, { "epoch": 0.7904, "grad_norm": 2.76302170753479, "learning_rate": 6.691819212762454e-06, "loss": 0.6323, "step": 247 }, { "epoch": 0.7936, "grad_norm": 3.0082130432128906, "learning_rate": 6.502640269896953e-06, "loss": 0.503, "step": 248 }, { "epoch": 0.7968, "grad_norm": 2.735734462738037, "learning_rate": 6.3157733509600355e-06, "loss": 0.5215, "step": 249 }, { "epoch": 0.8, "grad_norm": 3.3777284622192383, "learning_rate": 6.1312418128412565e-06, "loss": 0.5809, "step": 250 }, { "epoch": 0.8032, "grad_norm": 2.971823215484619, "learning_rate": 5.949068720525991e-06, "loss": 0.5922, "step": 251 }, { "epoch": 0.8064, "grad_norm": 2.5565667152404785, "learning_rate": 5.769276844212501e-06, "loss": 0.5451, "step": 252 }, { "epoch": 0.8096, "grad_norm": 2.3006818294525146, "learning_rate": 5.591888656465874e-06, "loss": 0.4815, "step": 253 }, { "epoch": 0.8128, "grad_norm": 3.275210380554199, "learning_rate": 5.416926329409083e-06, "loss": 0.5855, "step": 254 }, { "epoch": 0.816, "grad_norm": 2.7305386066436768, "learning_rate": 5.244411731951671e-06, "loss": 0.5612, "step": 255 }, { "epoch": 0.8192, "grad_norm": 2.7912111282348633, "learning_rate": 5.074366427056309e-06, "loss": 0.5502, "step": 256 }, { "epoch": 0.8224, "grad_norm": 2.3598272800445557, "learning_rate": 4.90681166904359e-06, "loss": 0.4626, "step": 257 }, { "epoch": 0.8256, "grad_norm": 2.215691566467285, "learning_rate": 4.741768400935417e-06, "loss": 0.5392, "step": 258 }, { "epoch": 0.8288, "grad_norm": 3.136962652206421, "learning_rate": 4.579257251837271e-06, "loss": 0.6118, "step": 259 }, { "epoch": 0.832, "grad_norm": 2.4404520988464355, "learning_rate": 4.419298534359759e-06, "loss": 0.5114, "step": 260 }, { "epoch": 0.8352, "grad_norm": 2.8911097049713135, "learning_rate": 4.261912242079674e-06, "loss": 0.5594, "step": 261 }, { "epoch": 0.8384, "grad_norm": 2.625516653060913, "learning_rate": 4.107118047040995e-06, "loss": 0.5123, "step": 262 }, { "epoch": 0.8416, "grad_norm": 3.0125982761383057, "learning_rate": 3.954935297295975e-06, "loss": 0.5648, "step": 263 }, { "epoch": 0.8448, "grad_norm": 2.3485519886016846, "learning_rate": 3.8053830144868547e-06, "loss": 0.5347, "step": 264 }, { "epoch": 0.848, "grad_norm": 3.10200572013855, "learning_rate": 3.6584798914682582e-06, "loss": 0.4884, "step": 265 }, { "epoch": 0.8512, "grad_norm": 2.7132773399353027, "learning_rate": 3.514244289970753e-06, "loss": 0.5918, "step": 266 }, { "epoch": 0.8544, "grad_norm": 2.857553005218506, "learning_rate": 3.3726942383057763e-06, "loss": 0.6052, "step": 267 }, { "epoch": 0.8576, "grad_norm": 2.5855906009674072, "learning_rate": 3.233847429112244e-06, "loss": 0.5894, "step": 268 }, { "epoch": 0.8608, "grad_norm": 2.1608877182006836, "learning_rate": 3.0977212171451e-06, "loss": 0.4059, "step": 269 }, { "epoch": 0.864, "grad_norm": 2.8646678924560547, "learning_rate": 2.9643326171061165e-06, "loss": 0.5117, "step": 270 }, { "epoch": 0.8672, "grad_norm": 3.1131443977355957, "learning_rate": 2.833698301517185e-06, "loss": 0.5713, "step": 271 }, { "epoch": 0.8704, "grad_norm": 2.588733434677124, "learning_rate": 2.7058345986363974e-06, "loss": 0.5429, "step": 272 }, { "epoch": 0.8736, "grad_norm": 3.296912908554077, "learning_rate": 2.5807574904171155e-06, "loss": 0.5944, "step": 273 }, { "epoch": 0.8768, "grad_norm": 2.6463258266448975, "learning_rate": 2.4584826105103764e-06, "loss": 0.5025, "step": 274 }, { "epoch": 0.88, "grad_norm": 2.728508710861206, "learning_rate": 2.3390252423108076e-06, "loss": 0.514, "step": 275 }, { "epoch": 0.8832, "grad_norm": 2.978658437728882, "learning_rate": 2.222400317046308e-06, "loss": 0.5277, "step": 276 }, { "epoch": 0.8864, "grad_norm": 2.75089955329895, "learning_rate": 2.108622411911773e-06, "loss": 0.5267, "step": 277 }, { "epoch": 0.8896, "grad_norm": 2.917853593826294, "learning_rate": 1.997705748247067e-06, "loss": 0.5621, "step": 278 }, { "epoch": 0.8928, "grad_norm": 2.567577838897705, "learning_rate": 1.8896641897594492e-06, "loss": 0.5322, "step": 279 }, { "epoch": 0.896, "grad_norm": 2.7284419536590576, "learning_rate": 1.78451124079074e-06, "loss": 0.5843, "step": 280 }, { "epoch": 0.8992, "grad_norm": 3.4555678367614746, "learning_rate": 1.6822600446293636e-06, "loss": 0.5511, "step": 281 }, { "epoch": 0.9024, "grad_norm": 2.928220272064209, "learning_rate": 1.5829233818675766e-06, "loss": 0.5303, "step": 282 }, { "epoch": 0.9056, "grad_norm": 2.5740914344787598, "learning_rate": 1.486513668803946e-06, "loss": 0.5162, "step": 283 }, { "epoch": 0.9088, "grad_norm": 3.3031325340270996, "learning_rate": 1.3930429558914494e-06, "loss": 0.5586, "step": 284 }, { "epoch": 0.912, "grad_norm": 3.0398950576782227, "learning_rate": 1.3025229262312366e-06, "loss": 0.5671, "step": 285 }, { "epoch": 0.9152, "grad_norm": 2.4170124530792236, "learning_rate": 1.214964894112361e-06, "loss": 0.4867, "step": 286 }, { "epoch": 0.9184, "grad_norm": 2.8461062908172607, "learning_rate": 1.1303798035975643e-06, "loss": 0.528, "step": 287 }, { "epoch": 0.9216, "grad_norm": 2.772353172302246, "learning_rate": 1.0487782271553504e-06, "loss": 0.5361, "step": 288 }, { "epoch": 0.9248, "grad_norm": 2.490363597869873, "learning_rate": 9.701703643385295e-07, "loss": 0.5397, "step": 289 }, { "epoch": 0.928, "grad_norm": 2.0275142192840576, "learning_rate": 8.94566040509337e-07, "loss": 0.4032, "step": 290 }, { "epoch": 0.9312, "grad_norm": 2.7005605697631836, "learning_rate": 8.219747056113586e-07, "loss": 0.528, "step": 291 }, { "epoch": 0.9344, "grad_norm": 2.321662187576294, "learning_rate": 7.524054329883346e-07, "loss": 0.4782, "step": 292 }, { "epoch": 0.9376, "grad_norm": 3.391035318374634, "learning_rate": 6.858669182500971e-07, "loss": 0.5802, "step": 293 }, { "epoch": 0.9408, "grad_norm": 2.873091697692871, "learning_rate": 6.223674781856592e-07, "loss": 0.6049, "step": 294 }, { "epoch": 0.944, "grad_norm": 3.041048288345337, "learning_rate": 5.619150497236992e-07, "loss": 0.5748, "step": 295 }, { "epoch": 0.9472, "grad_norm": 2.43678617477417, "learning_rate": 5.045171889404954e-07, "loss": 0.4379, "step": 296 }, { "epoch": 0.9504, "grad_norm": 2.705655336380005, "learning_rate": 4.501810701154907e-07, "loss": 0.4788, "step": 297 }, { "epoch": 0.9536, "grad_norm": 3.0620837211608887, "learning_rate": 3.98913484834551e-07, "loss": 0.5725, "step": 298 }, { "epoch": 0.9568, "grad_norm": 2.623995542526245, "learning_rate": 3.507208411410778e-07, "loss": 0.509, "step": 299 }, { "epoch": 0.96, "grad_norm": 3.280472755432129, "learning_rate": 3.0560916273504325e-07, "loss": 0.5453, "step": 300 }, { "epoch": 0.9632, "grad_norm": 2.764864683151245, "learning_rate": 2.635840882200924e-07, "loss": 0.4987, "step": 301 }, { "epoch": 0.9664, "grad_norm": 2.27424955368042, "learning_rate": 2.246508703987543e-07, "loss": 0.4541, "step": 302 }, { "epoch": 0.9696, "grad_norm": 2.5994205474853516, "learning_rate": 1.8881437561586722e-07, "loss": 0.5156, "step": 303 }, { "epoch": 0.9728, "grad_norm": 2.2249081134796143, "learning_rate": 1.5607908315035667e-07, "loss": 0.3998, "step": 304 }, { "epoch": 0.976, "grad_norm": 2.5577194690704346, "learning_rate": 1.264490846553279e-07, "loss": 0.47, "step": 305 }, { "epoch": 0.9792, "grad_norm": 2.661684036254883, "learning_rate": 9.992808364666373e-08, "loss": 0.5783, "step": 306 }, { "epoch": 0.9824, "grad_norm": 2.729504346847534, "learning_rate": 7.651939504010885e-08, "loss": 0.5448, "step": 307 }, { "epoch": 0.9856, "grad_norm": 2.986076831817627, "learning_rate": 5.622594473692067e-08, "loss": 0.5982, "step": 308 }, { "epoch": 0.9888, "grad_norm": 3.096761703491211, "learning_rate": 3.90502692581729e-08, "loss": 0.5652, "step": 309 }, { "epoch": 0.992, "grad_norm": 2.4944214820861816, "learning_rate": 2.4994515427695374e-08, "loss": 0.4978, "step": 310 }, { "epoch": 0.9952, "grad_norm": 2.4917850494384766, "learning_rate": 1.4060440103746964e-08, "loss": 0.537, "step": 311 }, { "epoch": 0.9984, "grad_norm": 2.160087823867798, "learning_rate": 6.249409959421803e-09, "loss": 0.4661, "step": 312 }, { "epoch": 1.0, "grad_norm": 4.158653259277344, "learning_rate": 1.5624013118137326e-09, "loss": 0.5729, "step": 313 } ], "logging_steps": 1, "max_steps": 313, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.2702536303312896e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }